Citation
Abstract
In this paper, we investigate the use of Explainable Artificial Intelligence (XAI) methods for the interpretation of two Convolutional Neural Network (CNN) classifiers in the field of remote sensing (RS). Specifically, the SegNet and Unet architectures for RS building information extraction and segmentation are evaluated using a comprehensive array of primary- and layer-attributions XAI methods. The attribution methods are quantitatively evaluated using the sensitivity metric. Based on the visualization of the different XAI methods, Deconvolution and GradCAM results in many of the study areas show reliability. Moreover, these methods are able to accurately interpret both Unet's and SegNet's decisions and managed to analyze and reveal the internal mechanisms in both models (confirmed by the low sensitivity scores). Overall, no single method stood out as the best one.
Download File
Full text not available from this repository.
Official URL or Download Paper: https://ieeexplore.ieee.org/document/10089337
|
Additional Metadata
Item Type: | Conference or Workshop Item (Paper) |
---|---|
Divisions: | Faculty of Computer Science and Information Technology |
DOI Number: | https://doi.org/10.1109/CSDE56538.2022.10089337 |
Publisher: | IEEE |
Keywords: | XAI method; Primary attribution; Layer attribution; GradCAM; IntegratedGradients; Explainability |
Depositing User: | Ms. Nuraida Ibrahim |
Date Deposited: | 05 Oct 2023 07:09 |
Last Modified: | 05 Oct 2023 07:09 |
Altmetrics: | http://www.altmetric.com/details.php?domain=psasir.upm.edu.my&doi=10.1109/CSDE56538.2022.10089337 |
URI: | http://psasir.upm.edu.my/id/eprint/37761 |
Statistic Details: | View Download Statistic |
Actions (login required)
View Item |