Transformer-based Land Use and Land Cover Classification with Explainability using Satellite Imagery
November 9, 2025 Β· View on GitHub
This repository contains the code for our paper:
π Transformer-based Land Use and Land Cover Classification with Explainability using Satellite Imagery
β Authors: Mehak Khan, Abdul Hanan, Meruyert Kenzhebay, Michele Gazzea & Reza Arghandeh
π Journal: Scientific Reports (Nature)
In this work, we introduce a framework that enhances the efficiency of Vision Transformer (ViT) and Swin Transformer models through transfer learning and fine-tuning techniques.
Our approach also emphasizes model interpretability, ensuring that deep learning decisions in Land Use and Land Cover (LULC) classification are both transparent and understandable. This is particularly crucial for forestry, agriculture, and environmental monitoring applications using satellite imagery.
π Key Features
β Transformer-based Deep Learning: Fine-tuned Vision Transformer (ViT) and Swin Transformer models for satellite image classification.
β Explainability with Integrated Gradients: We leverage Captumβs Integrated Gradients to provide interpretability in LULC classification.
β Efficient Training Pipeline: Utilizes transfer learning and fine-tuning for improved performance.
β Application Areas: Forestry, agricultural monitoring, environmental analysis, and urban planning.
π Dataset
We use the EuroSAT-RGB dataset, which contains RGB satellite images across ten different land use classes. For further validation of our frameworkβs generalization and scalability, we conducted additional experiments using PatternNet dataset.
Example images from EuroSAT:
π§ Models
Our framework leverages two transformer-based models:
- Vision Transformer (ViT)
- Swin Transformer
π Explainability
To ensure model interpretability, we integrate Integrated Gradients from the Captum Library. This allows us to visualize feature importance in the classification process.
π Acknowledgements
- The EuroSAT and PatternNet datasets are publicly available.
- We use Vision Transformers (ViT) and Swin Transformers, based on the Timm library library.
- Explainability is powered by the Captum Library.
π¬ Contact
For questions or collaborations, feel free to open an issue or reach out!
π§ Email: mehakkhan3@hotmail.com
π Citation
If you find this work useful, please cite our paper:
@article{khan2024transformer,
title={Transformer-based land use and land cover classification with explainability using satellite imagery},
author={Khan, Mehak and Hanan, Abdul and Kenzhebay, Meruyert and Gazzea, Michele and Arghandeh, Reza},
journal={Scientific Reports},
volume={14},
number={1},
pages={16744},
year={2024},
publisher={Nature Publishing Group UK London}
}