The browser you are using is not supported by this website. All versions of Internet Explorer are no longer supported, either by us or Microsoft (read more here: https://www.microsoft.com/en-us/microsoft-365/windows/end-of-ie-support).

Please use a modern browser to fully experience our website, such as the newest versions of Edge, Chrome, Firefox or Safari etc.

Photo of Mattias Ohlsson

Mattias Ohlsson

Professor

Photo of Mattias Ohlsson

A review of explainable AI in the satellite data, deep machine learning, and human poverty domain

Author

  • Ola Hall
  • Mattias Ohlsson
  • Thorsteinn Rögnvaldsson

Summary, in English

Recent advances in artificial intelligence and deep machine learning have created a step change in how to measure human development indicators, in particular asset-based poverty. The combination of satellite imagery and deep machine learning now has the capability to estimate some types of poverty at a level close to what is achieved with traditional household surveys. An increasingly important issue beyond static estimations is whether this technology can contribute to scientific discovery and, consequently, new knowledge in the poverty and welfare domain. A foundation for achieving scientific insights is domain knowledge, which in turn translates into explainability and scientific consistency. We perform an integrative literature review focusing on three core elements relevant in this context—transparency, interpretability, and explainability—and investigate how they relate to the poverty, machine learning, and satellite imagery nexus. Our inclusion criteria for papers are that they cover poverty/wealth prediction, using survey data as the basis for the ground truth poverty/wealth estimates, be applicable to both urban and rural settings, use satellite images as the basis for at least some of the inputs (features), and the method should include deep neural networks. Our review of 32 papers shows that the status of the three core elements of explainable machine learning (transparency, interpretability, and domain knowledge) is varied and does not completely fulfill the requirements set up for scientific insights and discoveries. We argue that explainability is essential to support wider dissemination and acceptance of this research in the development community and that explainability means more than just interpretability.

Department/s

  • Department of Human Geography
  • Computational Biology and Biological Physics - Undergoing reorganization
  • eSSENCE: The e-Science Collaboration

Publishing year

2022-10

Language

English

Publication/Series

Patterns

Volume

3

Issue

10

Document type

Journal article review

Publisher

Cell Press

Topic

  • Computer Science

Keywords

  • machine learning
  • poverty and satellite imagery
  • XAI

Status

Published

ISBN/ISSN/Other

  • ISSN: 2666-3899