Skip to main content

Lund University Publications

LUND UNIVERSITY LIBRARIES

Urban land use and land cover classification with interpretable machine learning – A case study using Sentinel-2 and auxiliary data

Hosseiny, Benyamin LU ; Abdi, Abdulhakim M. LU orcid and Jamali, Sadegh LU orcid (2022) In Remote Sensing Applications: Society and Environment 28.
Abstract
The European commission launch of the twin Sentinel-2 satellites provides new opportunities for land use and land cover (LULC) classification because of the readily availability of their data and their enhanced spatial, temporal and spectral resolutions. The rapid development of machine learning over the past decade led to data-driven models being at the forefront of high accuracy predictions of the physical world. However, the contribution of the driving variables behind these predictions cannot be explained beyond generalized metrics of overall performance. Here, we compared the performance of three shallow learners (support vector machines, random forest, and extreme gradient boosting) as well as two deep learners (a convolutional... (More)
The European commission launch of the twin Sentinel-2 satellites provides new opportunities for land use and land cover (LULC) classification because of the readily availability of their data and their enhanced spatial, temporal and spectral resolutions. The rapid development of machine learning over the past decade led to data-driven models being at the forefront of high accuracy predictions of the physical world. However, the contribution of the driving variables behind these predictions cannot be explained beyond generalized metrics of overall performance. Here, we compared the performance of three shallow learners (support vector machines, random forest, and extreme gradient boosting) as well as two deep learners (a convolutional neural network and a residual network with 50 layers) in and around the city of Malmö in southern Sweden. Our complete analysis suite involved 141 input features, 85 scenarios, and 8 LULC classes. We explored the interpretability of the five learners using Shapley additive explanations to better understand feature importance at the level of individual LULC classes. The purpose of class-level feature importance was to identify the most parsimonious combination of features that could reasonably map a particular class and enhance overall map accuracy. We showed that not only do overall accuracies increase from shallow (mean = 84.64%) to deep learners (mean = 92.63%) but that the number of explanatory variables required to obtain maximum accuracy decreases along the same gradient. Furthermore, we demonstrated that class-level importance metrics can be successfully identified using Shapley additive explanations in both shallow and deep learners, which allows for a more detailed understanding of variable importance. We show that for certain LULC classes there is a convergence of variable importance across all the algorithms, which helps explain model predictions and aid the selection of more parsimonious models. The use of class-level feature importance metrics is still new in LULC classification, and this study provides important insight into the potential of more nuanced importance metrics. (Less)
Please use this url to cite or link to this publication:
author
; and
organization
publishing date
type
Contribution to journal
publication status
published
subject
keywords
Remote sensing, Machine learning, Deep learning, Explainable AI, Earth observation, Sweden
in
Remote Sensing Applications: Society and Environment
volume
28
article number
100843
publisher
Elsevier
external identifiers
  • scopus:85139016934
ISSN
2352-9385
DOI
10.1016/j.rsase.2022.100843
language
English
LU publication?
yes
id
ed2a7530-3478-426d-b3f5-3ea2287f3bcd
date added to LUP
2022-09-29 14:54:06
date last changed
2023-05-10 11:36:37
@article{ed2a7530-3478-426d-b3f5-3ea2287f3bcd,
  abstract     = {{The European commission launch of the twin Sentinel-2 satellites provides new opportunities for land use and land cover (LULC) classification because of the readily availability of their data and their enhanced spatial, temporal and spectral resolutions. The rapid development of machine learning over the past decade led to data-driven models being at the forefront of high accuracy predictions of the physical world. However, the contribution of the driving variables behind these predictions cannot be explained beyond generalized metrics of overall performance. Here, we compared the performance of three shallow learners (support vector machines, random forest, and extreme gradient boosting) as well as two deep learners (a convolutional neural network and a residual network with 50 layers) in and around the city of Malmö in southern Sweden. Our complete analysis suite involved 141 input features, 85 scenarios, and 8 LULC classes. We explored the interpretability of the five learners using Shapley additive explanations to better understand feature importance at the level of individual LULC classes. The purpose of class-level feature importance was to identify the most parsimonious combination of features that could reasonably map a particular class and enhance overall map accuracy. We showed that not only do overall accuracies increase from shallow (mean = 84.64%) to deep learners (mean = 92.63%) but that the number of explanatory variables required to obtain maximum accuracy decreases along the same gradient. Furthermore, we demonstrated that class-level importance metrics can be successfully identified using Shapley additive explanations in both shallow and deep learners, which allows for a more detailed understanding of variable importance. We show that for certain LULC classes there is a convergence of variable importance across all the algorithms, which helps explain model predictions and aid the selection of more parsimonious models. The use of class-level feature importance metrics is still new in LULC classification, and this study provides important insight into the potential of more nuanced importance metrics.}},
  author       = {{Hosseiny, Benyamin and Abdi, Abdulhakim M. and Jamali, Sadegh}},
  issn         = {{2352-9385}},
  keywords     = {{Remote sensing; Machine learning; Deep learning; Explainable AI; Earth observation; Sweden}},
  language     = {{eng}},
  month        = {{09}},
  publisher    = {{Elsevier}},
  series       = {{Remote Sensing Applications: Society and Environment}},
  title        = {{Urban land use and land cover classification with interpretable machine learning – A case study using Sentinel-2 and auxiliary data}},
  url          = {{http://dx.doi.org/10.1016/j.rsase.2022.100843}},
  doi          = {{10.1016/j.rsase.2022.100843}},
  volume       = {{28}},
  year         = {{2022}},
}