Introduction
Deep learning has transformed the way satellite and aerial images are analyzed and interpreted. These images pose unique challenges, such as large sizes and diverse object classes, which offer opportunities for deep learning researchers. This repository offers a comprehensive overview of various deep learning techniques for analyzing satellite and aerial imagery, including architectures, models, and algorithms for tasks such as classification, segmentation, and object detection. It serves as a valuable resource for researchers, practitioners, and anyone interested in the latest advances in deep learning and its impact on computer vision and remote sensing. This repository is proudly sponsored by Orboculum
Orbuculum is an innovative and rapidly evolving platform designed with the specific intent to empower GIS and Earth Observation (EO) researchers by offering a unique avenue for monetizing their machine learning models. Standing distinctively apart from conventional marketplaces, Orbuculum pioneers a transformative approach by transmuting these models into smart contracts. This enables automatic remuneration for the creators each time their models are deployed, fostering an efficient and rewarding ecosystem.
Orbuculum's potential extends far beyond the reinvention of the GIS/EO research industry. It is poised to serve as an invaluable conduit for public welfare initiatives, especially those striving to mitigate climate change. By providing access to vital data and insightful analytics, Orbuculum promises to act as a potent resource in the ongoing battle against some of the most urgent global concerns. This integration of cutting-edge technology with socially impactful missions could position Orbuculum as an instrumental platform at the intersection of scientific research and sustainable development.
🖥️ Orbuculum website📺 Demo video of Orbuculum platform📺 Conversation between Robin and Derek Ding, the co-founder of the Orbuculum platform🗣️ Orbuculum Discord- 💻 Orbuculum Github
🐦 Orbuculum Twitter
How to use this repository: if you know exactly what you are looking for (e.g. you have the paper name) you can Control+F
to search for it in this page. Note BEGINNER
is used to identify material that is suitable for begginers & getting started with a topic
Techniques
- Classification
- Segmentation
- Instance segmentation
- Object detection
- Object counting
- Regression
- Cloud detection & removal
- Change detection
- Time series
- Crop classification
- Crop yield
- Wealth and economic activity
- Disaster response
- Super-resolution
- Pansharpening
- Image-to-image translation
- Data fusion
- Generative Adversarial Networks (GANs)
- Autoencoders, dimensionality reduction, image embeddings & similarity search
- Image retrieval
- Image Captioning
- Visual Question Answering
- Mixed data learning
- Few-shot learning
- Self-supervised, unsupervised & contrastive learning
- Weakly & semi-supervised learning
- Active learning
- Federated Learning
- Transformers
- Adversarial ML
- Image registration
- Terrain mapping, Disparity Estimation, Lidar, DEMs & NeRF
- Thermal Infrared
- SAR
- NDVI-Vegetation Index
- General image quality
- Synthetic data
1. Classification
The UC merced dataset is a well known classification dataset.
Classification is a fundamental task in remote sensing data analysis, where the goal is to assign a semantic label to each image, such as 'urban', 'forest', 'agricultural land', etc. The process of assigning labels to an image is known as image-level classification. However, in some cases, a single image might contain multiple different land cover types, such as a forest with a river running through it, or a city with both residential and commercial areas. In these cases, image-level classification becomes more complex and involves assigning multiple labels to a single image. This can be accomplished using a combination of feature extraction and machine learning algorithms to accurately identify the different land cover types. It is important to note that image-level classification should not be confused with pixel-level classification, also known as semantic segmentation. While image-level classification assigns a single label to an entire image, semantic segmentation assigns a label to each individual pixel in an image, resulting in a highly detailed and accurate representation of the land cover types in an image. Read A brief introduction to satellite image classification with neural networks
1.1. Land classification on Sentinel 2 data using a simple sklearn cluster algorithm or deep learning CNN BEGINNER
1.2. Land Use Classification on Merced dataset using CNN in Keras
or fastai. Also checkout Multi-label Land Cover Classification using the redesigned multi-label Merced dataset with 17 land cover classes BEGINNER
1.3. Multi-Label Classification of Satellite Photos of the Amazon Rainforest using keras or FastAI BEGINNER
1.4. EuroSat-Satellite-CNN-and-ResNet -> Classifying custom image datasets by creating Convolutional Neural Networks and Residual Networks from scratch with PyTorch BEGINNER
1.5. Detecting Informal Settlements from Satellite Imagery using fine-tuning of ResNet-50 classifier with repo
1.6. Land-Cover-Classification-using-Sentinel-2-Dataset -> well written Medium article accompanying this repo but using the EuroSAT dataset
1.7. Land Cover Classification of Satellite Imagery using Convolutional Neural Networks using Keras and a multi spectral dataset captured over vineyard fields of Salinas Valley, California
1.8. Detecting deforestation from satellite images -> using FastAI and ResNet50, with repo fsdl_deforestation_detection
1.9. Neural Network for Satellite Data Classification Using Tensorflow in Python -> A step-by-step guide for Landsat 5 multispectral data classification for binary built-up/non-built-up class prediction, with repo
1.10. Slums mapping from pretrained CNN network on VHR (Pleiades: 0.5m) and MR (Sentinel: 10m) imagery
1.11. Comparing urban environments using satellite imagery and convolutional neural networks -> includes interesting study of the image embedding features extracted for each image on the Urban Atlas dataset. Accompanying paper
1.12. RSI-CB -> A Large Scale Remote Sensing Image Classification Benchmark via Crowdsource Data. See also Remote-sensing-image-classification
1.13. NAIP_PoolDetection -> modelled as an object recognition problem, a CNN is used to identify images as being swimming pools or something else - specifically a street, rooftop, or lawn
1.14. Land Use and Land Cover Classification using a ResNet Deep Learning Architecture -> uses fastai and the EuroSAT dataset
1.15. Vision Transformers Use Case: Satellite Image Classification without CNNs
1.16. WaterNet -> a CNN that identifies water in satellite images
1.17. Road-Network-Classification -> Road network classification model using ResNet-34, road classes organic, gridiron, radial and no pattern
1.18. Scaling AI to map every school on the planet
1.19. Landsat classification CNN tutorial with repo
1.20. satellite-crosswalk-classification
1.22. Implementation of the 3D-CNN model for land cover classification -> uses the Sundarbans dataset, with repo. Also read Land cover classification of Sundarbans satellite imagery using K-Nearest Neighbor(K-NNC), Support Vector Machine (SVM), and Gradient Boosting classification algorithms which is by the same author and shares the repo
1.23. SSTN -> PyTorch Implementation of SSTNs for hyperspectral image classifications from the IEEE T-GRS paper "Spectral-Spatial Transformer Network for Hyperspectral Image Classification: A FAS Framework." Demonstrates a novel spectral-spatial transformer network (SSTN), which consists of spatial attention and spectral association modules, to overcome the constraints of convolution kernels
1.24. SatellitePollutionCNN -> A novel algorithm to predict air pollution levels with state-of-art accuracy using deep learning and GoogleMaps satellite images
1.25. PropertyClassification -> Classifying the type of property given Real Estate, satellite and Street view Images
1.26. remote-sense-quickstart -> classification on a number of datasets, including with attention visualization
1.27. Satellite image classification using multiple machine learning algorithms
1.28. satsense -> a Python library for land use/cover classification using classical features including HoG & NDVI
1.29. PyTorch_UCMerced_LandUse -> simple pytorch implementation fine tuned on ResNet and basic augmentations
1.30. EuroSAT-image-classification -> simple pytorch implementation fine tuned on ResNet
1.31. landcover_classification -> using fast.ai on EuroSAT
1.32. IGARSS2020_BWMS -> Band-Wise Multi-Scale CNN Architecture for Remote Sensing Image Scene Classification with a novel CNN architecture for the feature embedding of high-dimensional RS images
1.33. image.classification.on.EuroSAT -> solution in pure pytorch
1.34. hurricane_damage -> Post-hurricane structure damage assessment based on aerial imagery with CNN
1.35. openai-drivendata-challenge -> Using deep learning to classify the building material of rooftops (aerial imagery from South America)
1.36. is-it-abandoned -> Can we tell if a house is abandoned based on aerial LIDAR imagery?
1.37. BoulderAreaDetector -> CNN to classify whether a satellite image shows an area would be a good rock climbing spot or not
1.38. ISPRS_S2FL -> code for paper: Multimodal Remote Sensing Benchmark Datasets for Land Cover Classification with A Shared and Specific Feature Learning Model. S2FL is capable of decomposing multimodal RS data into modality-shared and modality-specific components, enabling the information blending of multi-modalities more effectively 1.39. Brazilian-Coffee-Detection -> uses Keras with public dataset
1.40. tf-crash-severity -> predict the crash severity for given road features contained within satellite images
1.41. ensemble_LCLU -> code for 2021 paper: Deep neural network ensembles for remote sensing land cover and land use classification
1.42. cerraNet -> contextually classify the types of use and coverage in the Brazilian Cerrado
1.43. Urban-Analysis-Using-Satellite-Imagery -> classify urban area as planned or unplanned using a combination of segmentation and classification
1.44. ChipClassification -> code for 2019 paper: Deep learning for multi-modal classification of cloud, shadow and land cover scenes in PlanetScope and Sentinel-2 imagery
1.45. DeeplearningClassficationLandsat-tImages -> Water/Ice/Land Classification Using Large-Scale Medium Resolution Landsat Satellite Images
1.46. wildfire-detection-from-satellite-images-ml -> detect whether an image contains a wildfire, with example flask web app
1.47. mining-discovery-with-deep-learning -> code for the 2020 paper: Mining and Tailings Dam Detection in Satellite Imagery Using Deep Learning
1.48. e-Farmerce-platform -> classify crop type
1.49. sentinel2-deep-learning -> Novel Training Methodologies for Land Classification of Sentinel-2 Imagery
1.50. RSSC-transfer -> code for 2021 paper: The Role of Pre-Training in High-Resolution Remote Sensing Scene Classification
1.51. Classifying Geo-Referenced Photos and Satellite Images for Supporting Terrain Classification -> detect floods
1.52. Pay-More-Attention -> code for 2021 paper: Remote Sensing Image Scene Classification Based on an Enhanced Attention Module
1.53. Remote-Sensing-Image-Classification-via-Improved-Cross-Entropy-Loss-and-Transfer-Learning-Strategy -> code for 2019 paper: Remote Sensing Image Classification via Improved Cross-Entropy Loss and Transfer Learning Strategy Based on Deep Convolutional Neural Networks
1.54. DenseNet40-for-HRRSISC -> DenseNet40 for remote sensing image scene classification, uses UC Merced Dataset
1.55. SKAL -> code for 2022 paper: Looking Closer at the Scene: Multiscale Representation Learning for Remote Sensing Image Scene Classification
1.56. potsdam-tensorflow-practice -> image classification of Potsdam dataset using tensorflow
1.57. SAFF -> code for 2021 paper: Self-Attention-Based Deep Feature Fusion for Remote Sensing Scene Classification
1.58. GLNET -> code for 2021 paper: Convolutional Neural Networks Based Remote Sensing Scene Classification under Clear and Cloudy Environments
1.59. Remote-sensing-image-classification -> transfer learning using pytorch to classify remote sensing data into three classes: aircrafts, ships, none
1.60. remote_sensing_pretrained_models -> as an alternative to fine tuning on models pretrained on ImageNet, here some CNN are pretrained on the RSD46-WHU & AID datasets
1.61. CNN_AircraftDetection -> CNN for aircraft detection in satellite images using keras
1.62. OBIC-GCN -> code for 2021 paper: Object-based Classification Framework of Remote Sensing Images with Graph Convolutional Networks
1.63. aitlas-arena -> An open-source benchmark framework for evaluating state-of-the-art deep learning approaches for image classification in Earth Observation (EO)
1.64. droughtwatch -> code for 2020 paper: Satellite-based Prediction of Forage Conditions for Livestock in Northern Kenya
1.65. JSTARS_2020_DPN-HRA -> code for 2020 paper: Deep Prototypical Networks With Hybrid Residual Attention for Hyperspectral Image Classification
1.66. SIGNA -> code for 2022 paper: Semantic Interleaving Global Channel Attention for Multilabel Remote Sensing Image Classification
1.67. Satellite Image Classification using rmldnn and Sentinel 2 data
1.68. PBDL -> code for 2022 paper: Patch-Based Discriminative Learning for Remote Sensing Scene Classification
1.69. EmergencyNet -> identify fire and other emergencies from a drone
1.70. satellite-deforestation -> Using Satellite Imagery to Identify the Leading Indicators of Deforestation, applied to the Kaggle Challenge Understanding the Amazon from Space
1.71. RSMLC -> code for 2023 paper: Deep Network Architectures as Feature Extractors for Multi-Label Classification of Remote Sensing Images
1.72. FireRisk -> A Remote Sensing Dataset for Fire Risk Assessment with Benchmarks Using Supervised and Self-supervised Learning
1.73. flood_susceptibility_mapping -> Towards urban flood susceptibility mapping using data-driven models in Berlin, Germany
1.74. tick-tick-bloom -> Winners of the Tick Tick Bloom: Harmful Algal Bloom Detection Challenge. Task was to predict severity of algae bloom, winners used decision trees
2. Segmentation
(left) a satellite image and (right) the semantic classes in the image.
Image segmentation is a crucial step in image analysis and computer vision, with the goal of dividing an image into semantically meaningful segments or regions. The process of image segmentation assigns a class label to each pixel in an image, effectively transforming an image from a 2D grid of pixels into a 2D grid of pixels with assigned class labels. One common application of image segmentation is road or building segmentation, where the goal is to identify and separate roads and buildings from other features within an image. To accomplish this task, single class models are often trained to differentiate between roads and background, or buildings and background. These models are designed to recognize specific features, such as color, texture, and shape, that are characteristic of roads or buildings, and use this information to assign class labels to the pixels in an image. Another common application of image segmentation is land use or crop type classification, where the goal is to identify and map different land cover types within an image. In this case, multi-class models are typically used to recognize and differentiate between multiple classes within an image, such as forests, urban areas, and agricultural land. These models are capable of recognizing complex relationships between different land cover types, allowing for a more comprehensive understanding of the image content. Read A brief introduction to satellite image segmentation with neural networks. Note that many articles which refer to 'hyperspectral land classification' are often actually describing semantic segmentation. Image source
2.1. Segmentation - Land use & land cover
2.1.1. U-Net for Semantic Segmentation on Unbalanced Aerial Imagery -> using the Dubai dataset BEGINNER
2.1.2. Semantic Segmentation of Dubai dataset Using a TensorFlow U-Net Model BEGINNER
2.1.3. nga-deep-learning -> performs semantic segmentation on high resultion GeoTIF data using a modified U-Net & Keras, published by NASA researchers
2.1.4. Automatic Detection of Landfill Using Deep Learning
2.1.5. SpectralNET -> a 2D wavelet CNN for Hyperspectral Image Classification, uses Salinas Scene dataset & Keras
2.1.6. laika -> The goal of this repo is to research potential sources of satellite image data and to implement various algorithms for satellite image segmentation
2.1.7. PEARL -> a human-in-the-loop AI tool to drastically reduce the time required to produce an accurate Land Use/Land Cover (LULC) map, blog post, uses Microsoft Planetary Computer and ML models run locally in the browser. Code for backelnd and frontend
2.1.8. Land Cover Classification with U-Net -> Satellite Image Multi-Class Semantic Segmentation Task with PyTorch Implementation of U-Net, uses DeepGlobe Land Cover Segmentation dataset, with code
2.1.9. Multi-class semantic segmentation of satellite images using U-Net using DSTL dataset, tensorflow 1 & python 2.7. Accompanying article
2.1.10. Codebase for multi class land cover classification with U-Net accompanying a masters thesis, uses Keras
2.1.11. dubai-satellite-imagery-segmentation -> due to the small dataset, image augmentation was used
2.1.12. CDL-Segmentation -> code for the 2021 paper: Deep Learning Based Land Cover and Crop Type Classification: A Comparative Study. Compares UNet, SegNet & DeepLabv3+
2.1.13. LoveDA -> code for the 2021 paper: A Remote Sensing Land-Cover Dataset for Domain Adaptive Semantic Segmentation
2.1.14. Satellite Imagery Semantic Segmentation with CNN -> 7 different segmentation classes, DeepGlobe Land Cover Classification Challenge dataset, with repo
2.1.15. Aerial Semantic Segmentation using U-Net Deep Learning Model medium article, with repo
2.1.16. UNet-Satellite-Image-Segmentation -> A Tensorflow implentation of light UNet semantic segmentation framework
2.1.17. DeepGlobe Land Cover Classification Challenge solution
2.1.18. Semantic-segmentation-with-PyTorch-Satellite-Imagery -> predict 25 classes on RGB imagery taken to assess the damage after Hurricane Harvey
2.1.19. Semantic Segmentation With Sentinel-2 Imagery -> uses LandCoverNet dataset and fast.ai
2.1.20. CNN_Enhanced_GCN -> code for 2021 paper: CNN-Enhanced Graph Convolutional Network With Pixel- and Superpixel-Level Feature Fusion for Hyperspectral Image Classification
2.1.21. LULCMapping-WV3images-CORINE-DLMethods -> Land Use and Land Cover Mapping Using Deep Learning Based Segmentation Approaches and VHR Worldview-3 Images
2.1.22. SOLC -> code for 2022 paper: MCANet: A joint semantic segmentation framework of optical and SAR images for land use classification. Uses WHU-OPT-SAR-dataset
2.1.23. MUnet-LUC -> Land Use with mUnet
2.1.24. land-cover -> code for 2021 paper: Model Generalization in Deep Learning Applications for Land Cover Mapping
2.1.25. generalizablersc -> code for 2022 paper: Cross-dataset Learning for Generalizable Land Use Scene Classification
2.1.26. Large-scale-Automatic-Identification-of-Urban-Vacant-Land -> code for 2022 paper: Large-scale automatic identification of urban vacant land using semantic segmentation of high-resolution remote sensing images
2.1.27. SSLTransformerRS -> code for 2022 paper: Self-supervised Vision Transformers for Land-cover Segmentation and Classification
2.1.28. aerial-tile-segmentation -> Large satellite image semantic segmentation into 6 classes using Tensorflow 2.0 and ISPRS benchmark dataset
2.1.29. LULCMapping-WV3images-CORINE-DLMethods -> code for 2022 paper: Land Use and Land Cover Mapping Using Deep Learning Based Segmentation Approaches and VHR Worldview-3 Images
2.1.30. DCSA-Net -> code for 2022 paper: Dynamic Convolution Self-Attention Network for Land-Cover Classification in VHR Remote-Sensing Images
2.1.31. CHeGCN-CNN_enhanced_Heterogeneous_Graph -> code for 2022 paper: CNN-Enhanced Heterogeneous Graph Convolutional Network: Inferring Land Use from Land Cover with a Case Study of Park Segmentation
2.1.32. TCSVT_2022_DGSSC -> code for the 2022 paper: DGSSC: A Deep Generative Spectral-Spatial Classifier for Imbalanced Hyperspectral Imagery
2.1.33. DeepForest-Wetland-Paper -> code for 2021 paper: Deep Forest classifier for wetland mapping using the combination of Sentinel-1 and Sentinel-2 data, GIScience & Remote Sensing
2.1.34. Wetland_UNet -> UNet models that can delineate wetlands using remote sensing data input including bands from Sentinel-2 LiDAR and geomorphons. By the Conservation Innovation Center of Chesapeake Conservancy and Defenders of Wildlife
2.1.35. DeepGlobe2018 -> PyTorch U-net for multi-class semantic segmentation
2.1.36. dynamicworld -> code for 2022 paper: Dynamic World, Near real-time global 10 m land use land cover mapping
2.1.37. spada -> code for 2023 paper: Land Cover Segmentation with Sparse Annotations from Sentinel-2 Imagery
2.2. Segmentation - Vegetation, crops & crop boundaries
2.2.1. Сrор field boundary detection: approaches and main challenges -> Medium article, covering historical and modern approaches BEGINNER
2.2.2. kenya-crop-mask -> Annual and in-season crop mapping in Kenya - LSTM classifier to classify pixels as containing crop or not, and a multi-spectral forecaster that provides a 12 month time series given a partial input. Dataset downloaded from GEE and pytorch lightning used for training BEGINNER
2.2.3. What’s growing there? Identify crops from multi-spectral remote sensing data (Sentinel 2) using eo-learn for data pre-processing, cloud detection, NDVI calculation, image augmentation & fastai
2.2.4. Tree species classification from from airborne LiDAR and hyperspectral data using 3D convolutional neural networks accompanies research paper and uses fastai
2.2.5. crop-type-classification -> using Sentinel 1 & 2 data with a U-Net + LSTM, more features (i.e. bands) and higher resolution produced better results (article, no code)
2.2.6. Find sports fields using Mask R-CNN and overlay on open-street-map
2.2.7. An LSTM to generate a crop mask for Togo
2.2.8. DeepSatModels -> Code for paper "Context-self contrastive pretraining for crop type semantic segmentation"
2.2.9. farm-pin-crop-detection-challenge -> Using eo-learn and fastai to identify crops from multi-spectral remote sensing data
2.2.10. Detecting Agricultural Croplands from Sentinel-2 Satellite Imagery -> We developed UNet-Agri, a benchmark machine learning model that classifies croplands using open-access Sentinel-2 imagery at 10m spatial resolution
2.2.11. DeepTreeAttention -> Implementation of Hang et al. 2020 "Hyperspectral Image Classification with Attention Aided CNNs" for tree species prediction
2.2.12. Crop-Classification -> crop classification using multi temporal satellite images
2.2.13. ParcelDelineation -> using a French polygons dataset and unet in keras
2.2.14. crop-mask -> End-to-end workflow for generating high resolution cropland maps, uses GEE & LSTM model
2.2.15. DeepCropMapping -> A multi-temporal deep learning approach with improved spatial generalizability for dynamic corn and soybean mapping, uses LSTM
2.2.16. Segment Canopy Cover and Soil using NDVI and Rasterio
2.2.17. Use KMeans clustering to segment satellite imagery by land cover/land use
2.2.18. ResUnet-a -> Implementation of the paper "ResUNet-a: a deep learning framework for semantic segmentation of remotely sensed data" in TensorFlow
2.2.19. DSD_paper_2020 -> The code for the paper: Crop Type Classification based on Machine Learning with Multitemporal Sentinel-1 Data
2.2.20. MR-DNN -> extract rice field from Landsat 8 satellite imagery
2.2.21. deep_learning_forest_monitoring -> Estimate vegetation height, code for paper: Forest mapping and monitoring of the African continent using Sentinel-2 data and deep learning
2.2.22. global-cropland-mapping -> global multi-temporal cropland mapping
2.2.23. U-Net for Semantic Segmentation of Soyabean Crop Fields with SAR images
2.2.24. UNet-RemoteSensing -> uses 7 bands of Landsat and keras
2.2.25. Landuse_DL -> delineate landforms due to the thawing of ice-rich permafrost
2.2.26. canopy -> code for 2019 paper: A Convolutional Neural Network Classifier Identifies Tree Species in Mixed-Conifer Forest from Hyperspectral Imagery
2.2.27. RandomForest-Classification -> script is for random forest classification of remote sensing multi-band images, used in 2019 paper: Multisensor data to derive peatland vegetation communities using a fixed-wing unmanned aerial vehicle
2.2.28. forest_change_detection -> forest change segmentation with time-dependent models, including Siamese, UNet-LSTM, UNet-diff, UNet3D models. Code for 2021 paper: Deep Learning for Regular Change Detection in Ukrainian Forest Ecosystem With Sentinel-2
2.2.29. cultionet -> segmentation of cultivated land, built on PyTorch Geometric and PyTorch Lightning
2.2.30. sentinel-tree-cover -> code for 2020 paper: A global method to identify trees outside of closed-canopy forests with medium-resolution satellite imagery
2.2.31. crop-type-detection-ICLR-2020 -> Winning Solutions from Crop Type Detection Competition at CV4A workshop, ICLR 2020
2.2.32. Crop identification using satellite imagery -> Medium article, introduction to crop identification
2.2.33. S4A-Models -> Various experiments on the Sen4AgriNet dataset
2.2.34. attention-mechanism-unet -> code for 2022 paper: An attention-based U-Net for detecting deforestation within satellite sensor imagery
2.2.35. Cocoa_plantations_detection -> Detecting cocoa plantation in Ivory Coast using Sentinel-2 remote sensing data using KNN, SVM, Random Forest and MLP
2.2.36. SummerCrop_Deeplearning -> code for 2022 paper: A Transferable Learning Classification Model and Carbon Sequestration Estimation of Crops in Farmland Ecosystem
2.2.37. DeepForest is a python package for training and predicting individual tree crowns from airborne RGB imagery
2.2.38. Official repository for the "Identifying trees on satellite images" challenge from Omdena
2.2.39. Counting-Trees-using-Satellite-Images -> create an inventory of incoming and outgoing trees for an annual tree inspections, uses keras & semantic segmentation
2.2.40. 2020 Nature paper - An unexpectedly large count of trees in the West African Sahara and Sahel -> tree detection framework based on U-Net & tensorflow 2 with code here
2.2.41. TreeDetection -> A color-based classifier to detect the trees in google image data along with tree visual localization and crown size calculations via OpenCV
2.2.42. PTDM -> code for 2022 paper: Pomelo Tree Detection Method Based on Attention Mechanism and Cross-Layer Feature Fusion
2.2.43. urban-tree-detection -> code for 2022 paper: Individual Tree Detection in Large-Scale Urban Environments using High-Resolution Multispectral Imagery. With dataset
2.2.44. BioMassters_baseline -> a basic pytorch lightning baseline using a UNet for getting started with the BioMassters challenge (biomass estimation)
2.2.45. Biomassters winners -> top 3 solutions
2.2.46. kbrodt biomassters solution -> 1st place solution
2.2.47. quqixun biomassters solution
2.2.48. biomass-estimation -> from Azavea, applied to Sentinel 1 & 2
2.2.49. 3DUNetGSFormer -> code for 2022 paper: 3DUNetGSFormer: A deep learning pipeline for complex wetland mapping using generative adversarial networks and Swin transformer
2.2.50. SEANet_torch -> code for 2023 paper: Using a semantic edge-aware multi-task neural network to delineate agricultural parcels from remote sensing images
2.2.51. arborizer -> Tree crowns segmentation and classification
2.2.52. ReUse -> UNet to estimate carbon absorbed by forests, using Biomass & Sentinel-2 imagery. Code for paper: ReUse: REgressive Unet for Carbon Storage and Above-Ground Biomass Estimation
2.2.53. unet-sentinel -> UNet to handle Sentinel-1 SAR images to identify deforestation
2.2.54. MaskedSST -> code for 2023 paper: Masked Vision Transformers for Hyperspectral Image Classification
2.3. Segmentation - Water, coastlines & floods
2.3.1. pytorch-waterbody-segmentation -> UNET model trained on the Satellite Images of Water Bodies dataset from Kaggle. The model is deployed on Hugging Face Spaces BEGINNER
2.3.2. Flood Detection and Analysis using UNET with Resnet-34 as the back bone uses fastai BEGINNER
2.3.3. Automatic Flood Detection from Satellite Images Using Deep Learning BEGINNER
2.3.4. UNSOAT used fastai to train a Unet to perform semantic segmentation on satellite imageries to detect water - paper + notebook, accuracy 0.97, precision 0.91, recall 0.92
2.3.6. Houston_flooding -> labeling each pixel as either flooded or not using data from Hurricane Harvey. Dataset consisted of pre and post flood images, and a ground truth floodwater mask was created using unsupervised clustering (with DBScan) of image pixels with human cluster verification/adjustment
2.3.7. ml4floods -> An ecosystem of data, models and code pipelines to tackle flooding with ML
2.3.8. A comprehensive guide to getting started with the ETCI Flood Detection competition -> using Sentinel1 SAR & pytorch
2.3.9. Map Floodwater of SAR Imagery with SageMaker -> applied to Sentinel-1 dataset
2.3.10. 1st place solution for STAC Overflow: Map Floodwater from Radar Imagery hosted by Microsoft AI for Earth -> combines Unet with Catboostclassifier, taking their maxima, not the average
2.3.11. hydra-floods -> an open source Python application for downloading, processing, and delivering surface water maps derived from remote sensing data
2.3.12. CoastSat -> tool for mapping coastlines which has an extension CoastSeg using segmentation models
2.3.13. Satellite_Flood_Segmentation_of_Harvey -> explores both deep learning and traditional kmeans
2.3.14. Flood Event Detection Utilizing Satellite Images
2.3.15. ETCI-2021-Competition-on-Flood-Detection -> Experiments on Flood Segmentation on Sentinel-1 SAR Imagery with Cyclical Pseudo Labeling and Noisy Student Training, with arxiv paper
2.3.16. FDSI -> Flood Detection in Satellite Images - 2017 Multimedia Satellite Task
2.3.17. deepwatermap -> a deep model that segments water on multispectral images
2.3.18. rivamap -> an automated river analysis and mapping engine
2.3.19. deep-water -> track changes in water level
2.3.20. WatNet -> A deep ConvNet for surface water mapping based on Sentinel-2 image, uses the Earth Surface Water Dataset
2.3.21. A-U-Net-for-Flood-Extent-Mapping -> in keras
2.3.22. floatingobjects -> code for the paper: TOWARDS DETECTING FLOATING OBJECTS ON A GLOBAL SCALE WITHLEARNED SPATIAL FEATURES USING SENTINEL 2. Uses U-Net & pytorch
2.3.23. SpaceNet8 -> baseline Unet solution to detect flooded roads and buildings
2.3.24. dlsim -> code for 2020 paper: Breaking the Limits of Remote Sensing by Simulation and Deep Learning for Flood and Debris Flow Mapping
2.3.25. Water-HRNet -> HRNet trained on Sentinel 2
2.3.26. semantic segmentation model to identify newly developed or flooded land using NAIP imagery provided by the Chesapeake Conservancy, training on MS Azure
2.3.27. BandNet -> code for 2022 paper: Analysis and application of multispectral data for water segmentation using machine learning. Uses Sentinel-2 data
2.3.28. mmflood -> code for 2022 paper: MMFlood: A Multimodal Dataset for Flood Delineation From Satellite Imagery (Sentinel 1 SAR)
2.3.29. Urban_flooding -> Towards transferable data-driven models to predict urban pluvial flood water depth in Berlin, Germany
2.3.30. Flood-Mapping-Using-Satellite-Images -> masters thesis comparing Random Forest & Unet
2.4. Segmentation - Fire, smoke & burn areas
2.4.1. SatelliteVu-AWS-Disaster-Response-Hackathon -> fire spread prediction using classical ML & deep learning BEGINNER
2.4.2. Wild Fire Detection using U-Net trained on Databricks & Keras, semantic segmentation
2.4.3. A Practical Method for High-Resolution Burned Area Monitoring Using Sentinel-2 and VIIRS with code. Dataset created on Google Earth Engine, downloaded to local machine for model training using fastai. The BA-Net model used is much smaller than U-Net, resulting in lower memory requirements and a faster computation
2.4.4. AI Geospatial Wildfire Risk Prediction -> A predictive model using geospatial raster data to asses wildfire hazard potential over the contiguous United States using Unet
2.4.5. IndustrialSmokePlumeDetection -> using Sentinel-2 & a modified ResNet-50
2.4.6. burned-area-detection -> uses Sentinel-2
2.4.7. rescue -> code of the paper: Attention to fires: multi-channel deep-learning models forwildfire severity prediction
2.4.8. smoke_segmentation -> Segmenting smoke plumes and predicting density from GOES imagery
2.4.9. wildfire-detection -> Using Vision Transformers for enhanced wildfire detection in satellite images
2.4.10. Burned_Area_Detection -> Detecting Burned Areas with Sentinel-2 data
2.4.11. burned-area-baseline -> baseline unet model accompanying the Satellite Burned Area Dataset (Sentinel 1 & 2)
2.4.12. burned-area-seg -> Burned area segmentation from Sentinel-2 using multi-task learning
2.5. Segmentation - Landslides
2.5.1. landslide4sense -> a competition focused on landslide detection using globally distributed multi-source satellite imagery. Baseline solution unet BEGINNER
2.5.2. landslide-mapping-with-cnn -> code for 2021 paper: A new strategy to map landslides with a generalized convolutional neural network
2.5.3. Relict_landslides_CNN_kmeans -> code for 2022 paper: Relict landslide detection in rainforest areas using a combination of k-means clustering algorithm and Deep-Learning semantic segmentation models
2.5.4. Landslide-mapping-on-SAR-data-by-Attention-U-Net -> code for 2022 paper: Rapid Mapping of landslide on SAR data by Attention U-net
2.5.5. SAR-landslide-detection-pretraining -> code for the 2022 paper: SAR-based landslide classification pretraining leads to better segmentation
2.5.6. landslide-sar-unet -> code for 2022 paper: Deep Learning for Rapid Landslide Detection using Synthetic Aperture Radar (SAR) Datacubes
2.6. Segmentation - Glaciers
2.6.1. HED-UNet -> a model for simultaneous semantic segmentation and edge detection, examples provided are glacier fronts and building footprints using the Inria Aerial Image Labeling dataset
2.6.2. glacier_mapping -> Mapping glaciers in the Hindu Kush Himalaya, Landsat 7 images, Shapefile labels of the glaciers, Unet with dropout
2.6.3. glacier-detect-ML -> a simple logistic regression model to identify a glacier in Landsat satellite imagery
2.6.4. GlacierSemanticSegmentation -> uses unet
2.6.5. Antarctic-fracture-detection -> uses UNet with the MODIS Mosaic of Antarctica to detect surface fractures (paper)
2.7. Segmentation - Other environmental
2.7.1. Detection of Open Landfills -> uses Sentinel-2 to detect large changes in the Normalized Burn Ratio (NBR)
2.7.2. sea_ice_remote_sensing -> Sea Ice Concentration classification
2.7.3. Methane-detection-from-hyperspectral-imagery -> code for 2020 paper: Deep Remote Sensing Methods for Methane Detection in Overhead Hyperspectral Imagery
2.7.4. EddyNet -> A Deep Neural Network For Pixel-Wise Classification of Oceanic Eddies
2.7.5. schisto-vegetation -> code for 2022 paper: Deep Learning Segmentation of Satellite Imagery Identifies Aquatic Vegetation Associated with Snail Intermediate Hosts of Schistosomiasis in Senegal, Africa
2.7.6. earth-forecasting-transformer -> code for 2022 paper: Earthformer: exploring space-time transformers for earth system forecasting
2.7.7. weather4cast-2022 -> Unet-3D baseline model for Weather4cast Rain Movie Prediction competition
2.7.8. WeatherFusionNet -> code for paper: WeatherFusionNet: Predicting Precipitation from Satellite Data. weather4cast-2022 1st place solution
2.7.9. marinedebrisdetector -> code for paper: Large-scale Detection of Marine Debris in Coastal Areas with Sentinel-2
2.8. Segmentation - Roads
Extracting roads is challenging due to the occlusions caused by other objects and the complex traffic environment
2.8.1. Road detection using semantic segmentation and albumentations for data augmention using the Massachusetts Roads Dataset, U-net & Keras. With code BEGINNER
2.8.2. ML_EPFL_Project_2 -> U-Net in Pytorch to perform semantic segmentation of roads on satellite images BEGINNER
2.8.3. Semantic Segmentation of roads using U-net Keras, OSM data, project summary article by student, no code
2.8.4. Winning Solutions from SpaceNet Road Detection and Routing Challenge
2.8.5. RoadVecNet -> Road-Network-Segmentation-and-Vectorization in keras with dataset and paper
2.8.6. Detecting road and road types jupyter notebook
2.8.7. awesome-deep-map -> A curated list of resources dedicated to deep learning / computer vision algorithms for mapping. The mapping problems include road network inference, building footprint extraction, etc.
2.8.8. RoadTracer: Automatic Extraction of Road Networks from Aerial Images -> uses an iterative search process guided by a CNN-based decision function to derive the road network graph directly from the output of the CNN
2.8.9. road_detection_mtl -> Road Detection using a multi-task Learning technique to improve the performance of the road detection task by incorporating prior knowledge constraints, uses the SpaceNet Roads Dataset
2.8.10. road_connectivity -> Improved Road Connectivity by Joint Learning of Orientation and Segmentation (CVPR2019)
2.8.11. Road-Network-Extraction using classical Image processing -> blur & canny edge detection
2.8.12. SPIN_RoadMapper -> Extracting Roads from Aerial Images via Spatial and Interaction Space Graph Reasoning for Autonomous Driving
2.8.13. road_extraction_remote_sensing -> pytorch implementation, CVPR2018 DeepGlobe Road Extraction Challenge submission. See also DeepGlobe-Road-Extraction-Challenge
2.8.14. RoadDetections dataset by Microsoft
2.8.15. CoANet -> Connectivity Attention Network for Road Extraction From Satellite Imagery. The CoA module incorporates graphical information to ensure the connectivity of roads are better preserved. With paper
2.8.16. Satellite Imagery Road Segmentation -> intro articule on Medium using the kaggle Massachusetts Roads Dataset
2.8.17. Label-Pixels -> for semantic segmentation of roads and other features
2.8.18. Satellite-image-road-extraction -> code for 2018 paper: Road Extraction by Deep Residual U-Net
2.8.19. road_building_extraction -> Pytorch implementation of U-Net architecture for road and building extraction
2.8.20. Satellite-Imagery-Road-Extraction -> research project in keras
2.8.21. SGCN -> code for 2021 paper: Split Depth-Wise Separable Graph-Convolution Network for Road Extraction in Complex Environments From High-Resolution Remote-Sensing Images
2.8.22. ASPN -> code for 2020 paper: Road Segmentation for Remote Sensing Images using Adversarial Spatial Pyramid Networks
2.8.23. FCNs-for-road-extraction-keras -> Road extraction of high-resolution remote sensing images based on various semantic segmentation networks
2.8.24. cresi -> Road network extraction from satellite imagery, with speed and travel time estimates
2.8.25. road-extraction-d-linknet -> code for 2018 paper: D-LinkNet: LinkNet with Pretrained Encoder and Dilated Convolution for High Resolution Satellite Imagery Road Extraction
2.8.26. Sat2Graph -> code for 2020 paper: Road Graph Extraction through Graph-Tensor Encoding
2.8.27. Image-Segmentation) -> using Massachusetts Road dataset and fast.ai
2.8.28. RoadTracer-M -> code for 2019 paper: Road Network Extraction from Satellite Images Using CNN Based Segmentation and Tracing
2.8.29. ScRoadExtractor -> code for 2020 paper: Scribble-based Weakly Supervised Deep Learning for Road Surface Extraction from Remote Sensing Images
2.8.30. RoadDA -> code for 2021 paper: Stagewise Unsupervised Domain Adaptation with Adversarial Self-Training for Road Segmentation of Remote Sensing Images
2.8.31. DeepSegmentor -> A Pytorch implementation of DeepCrack and RoadNet projects
2.8.32. Cascade_Residual_Attention_Enhanced_for_Refinement_Road_Extraction -> code for 2021 paper: Cascaded Residual Attention Enhanced Road Extraction from Remote Sensing Images
2.8.33. nia-road-baseline -> code for 2020 paper: NL-LinkNet: Toward Lighter but More Accurate Road Extraction with Non-Local Operations
2.8.34. IRSR-net -> code for 2022 paper: Lightweight Remote Sensing Road Detection Network
2.8.35. hironex -> A python tool for automatic, fully unsupervised extraction of historical road networks from historical maps
2.8.36. Road_detection_model -> code for 2022 paper: Mapping Roads in the Brazilian Amazon with Artificial Intelligence and Sentinel-2
2.8.37. DTnet -> code for 2022 paper: Road detection via a dual-task network based on cross-layer graph fusion modules
2.8.38. Automatic-Road-Extraction-from-Historical-Maps-using-Deep-Learning-Techniques -> code for the paper: Automatic Road Extraction from Historical Maps using Deep Learning Techniques: A Regional Case Study of Turkey in a German World War II map
2.8.39. Istanbul_Dataset -> segmentation on the Istanbul, Inria and Massachusetts datasets
2.8.40. Road-Segmentation -> Road segmentation on Satellite Images using CNN (U-Nets and FCN8) and Logistic Regression
2.8.41. D-LinkNet -> 1st place solution in DeepGlobe Road Extraction Challenge
2.8.42. PaRK-Detect -> code for 2023 paper: PaRK-Detect: Towards Efficient Multi-Task Satellite Imagery Road Extraction via Patch-Wise Keypoints Detection
2.8.43. tile2net -> code for 2023 paper: Mapping the walk: A scalable computer vision approach for generating sidewalk network datasets from aerial imagery
2.9. Segmentation - Buildings & rooftops
2.9.1. Road and Building Semantic Segmentation in Satellite Imagery uses U-Net on the Massachusetts Roads Dataset & keras BEGINNER
2.9.2. find-unauthorized-constructions-using-aerial-photography -> semantic segmentation using U-Net with custom_f1 metric & Keras. The creation of the dataset is described in this article BEGINNER
2.9.3. Semantic Segmentation on Aerial Images using fastai uses U-Net on the Inria Aerial Image Labeling Dataset of urban settlements in Europe and the United States, and is labelled as a building and not building classes (no repo) BEGINNER
2.9.4. Building footprint detection with fastai on the challenging SpaceNet7 dataset uses U-Net & fastai BEGINNER
2.9.5. Pix2Pix-for-Semantic-Segmentation-of-Satellite-Images -> using Pix2Pix GAN network to segment the building footprint from Satellite Images, uses tensorflow
2.9.6. SpaceNetUnet -> Baseline model is U-net like, applied to SpaceNet Vegas data, using Keras
2.9.7. automated-building-detection -> Input: very-high-resolution (<= 0.5 m/pixel) RGB satellite images. Output: buildings in vector format (geojson), to be used in digital map products. Built on top of robosat and robosat.pink.
2.9.8. project_sunroof_india -> Analyzed Google Satellite images to generate a report on individual house rooftop's solar power potential, uses a range of classical computer vision techniques (e.g Canny Edge Detection) to segment the roofs
2.9.9. JointNet-A-Common-Neural-Network-for-Road-and-Building-Extraction
2.9.10. Mapping Africa’s Buildings with Satellite Imagery: Google AI blog post. See the open-buildings dataset
2.9.11. nz_convnet -> A U-net based ConvNet for New Zealand imagery to classify building outlines
2.9.12. polycnn -> End-to-End Learning of Polygons for Remote Sensing Image Classification
2.9.13. spacenet_building_detection solution by motokimura using Unet
2.9.14. How to extract building footprints from satellite images using deep learning
2.9.15. Vec2Instance -> applied to the SpaceNet challenge AOI 2 (Vegas) building footprint dataset, tensorflow v1.12
2.9.16. EarthquakeDamageDetection -> Buildings segmentation from satellite imagery and damage classification for each build, using Keras
2.9.17. Semantic-segmentation repo by fuweifu-vtoo -> uses pytorch and the Massachusetts Buildings & Roads Datasets
2.9.18. Extracting buildings and roads from AWS Open Data using Amazon SageMaker -> uses merged RGB (SpaceNet) and LiDAR (USGS 3DEP) datasets with Unet to reproduce the winning algorithm from SpaceNet challenge 4 by XD_XD. With repo
2.9.19. TF-SegNet -> AirNet is a segmentation network based on SegNet, but with some modifications
2.9.20. rgb-footprint-extract -> a Semantic Segmentation Network for Urban-Scale Building Footprint Extraction Using RGB Satellite Imagery, DeepLavV3+ module with a Dilated ResNet C42 backbone
2.9.21. SpaceNetExploration -> A sample project demonstrating how to extract building footprints from satellite images using a semantic segmentation model. Data from the SpaceNet Challenge
2.9.22. Rooftop-Instance-Segmentation -> VGG-16, Instance Segmentation, uses the Airs dataset
2.9.23. solar-farms-mapping -> An Artificial Intelligence Dataset for Solar Energy Locations in India
2.9.24. poultry-cafos -> This repo contains code for detecting poultry barns from high-resolution aerial imagery and an accompanying dataset of predicted barns over the United States
2.9.25. ssai-cnn -> This is an implementation of Volodymyr Mnih's dissertation methods on his Massachusetts road & building dataset
2.9.26. Remote-sensing-building-extraction-to-3D-model-using-Paddle-and-Grasshopper
2.9.27. segmentation-enhanced-resunet -> Urban building extraction in Daejeon region using Modified Residual U-Net (Modified ResUnet) and applying post-processing
2.9.28. Mask RCNN for Spacenet Off Nadir Building Detection
2.9.29. GRSL_BFE_MA -> Deep Learning-based Building Footprint Extraction with Missing Annotations using a novel loss function
2.9.30. FER-CNN -> Detection, Classification and Boundary Regularization of Buildings in Satellite Imagery Using Faster Edge Region Convolutional Neural Networks, with paper
2.9.31. UNET-Image-Segmentation-Satellite-Picture -> Unet to predict roof tops on Crowed AI Mapping dataset, uses keras
2.9.32. Vector-Map-Generation-from-Aerial-Imagery-using-Deep-Learning-GeoSpatial-UNET -> applied to geo-referenced images which are very large size > 10k x 10k pixels
2.9.33. building-footprint-segmentation -> pip installable library to train building footprint segmentation on satellite and aerial imagery, applied to Massachusetts Buildings Dataset and Inria Aerial Image Labeling Dataset
2.9.34. SemSegBuildings -> Project using fast.ai framework for semantic segmentation on Inria building segmentation dataset
2.9.35. FCNN-example -> overfit to a given single image to detect houses
2.9.36. SAT2LOD2 -> an open-source, python-based GUI-enabled software that takes the satellite images as inputs and returns LoD2 building models as outputs, with paper
2.9.37. SatFootprint -> building segmentation on the Spacenet 7 dataset
2.9.38. Building-Detection -> code for running a Raster Vision experiment to train a model to detect buildings from satellite imagery in three cities in Latin America
2.9.39. Multi-building-tracker -> code for paper: Multi-target building tracker for satellite images using deep learning
2.9.40. Boundary Enhancement Semantic Segmentation for Building Extraction
2.9.41. UNet_keras_for_RSimage -> keras code for binary semantic segmentation
2.9.42. Spacenet-Building-Detection -> uses keras
2.9.43. LGPNet-BCD -> code for 2021 paper: Building Change Detection for VHR Remote Sensing Images via Local-Global Pyramid Network and Cross-Task Transfer Learning Strategy
2.9.44. MTL_homoscedastic_SRB -> code for 2021 paper: A Multi-Task Deep Learning Framework for Building Footprint Segmentation
2.9.45. UNet_CNN -> UNet model to segment building coverage in Boston using Remote sensing data, uses keras
2.9.46. FDANet -> code for 2021 paper: Full-Level Domain Adaptation for Building Extraction in Very-High-Resolution Optical Remote-Sensing Images
2.9.47. CBRNet -> code for 2022 paper: A Coarse-to-fine Boundary Refinement Network for Building Extraction from Remote Sensing Imagery
2.9.48. ASLNet -> code for 2021 paper: Adversarial Shape Learning for Building Extraction in VHR Remote Sensing Images
2.9.49. BRRNet -> implementation of Modified U-Net from 2020 paper: BRRNet: A Fully Convolutional Neural Network for Automatic Building Extraction From High-Resolution Remote Sensing Images
2.9.50. Multi-Scale-Filtering-Building-Index -> Python implementation of building extraction index proposed in 2019 paper: A Multi - Scale Filtering Building Index for Building Extraction in Very High - Resolution Satellite Imagery
2.9.51. Models for Remote Sensing -> long list of unets etc applied to building detection
2.9.52. boundary_loss_for_remote_sensing -> code for 2019 paper: Boundary Loss for Remote Sensing Imagery Semantic Segmentation
2.9.53. Open Cities AI Challenge -> Segmenting Buildings for Disaster Resilience. Winning solutions on Github
2.9.54. MAPNet -> code for 2020 paper: Multi Attending Path Neural Network for Building Footprint Extraction from Remote Sensed Imagery
2.9.55. dual-hrnet -> localizing buildings and classifying their damage level
2.9.56. ESFNet -> code for 2019 paper: Efficient Network for Building Extraction from High-Resolution Aerial Images
2.9.57. rooftop-detection-python -> Detect Rooftops from low resolution satellite images and calculate area for cultivation and solar panel installment using classical computer vision techniques
2.9.58. keras_segmentation_models -> code for 2022 paper: Using Open Vector-Based Spatial Data to Create Semantic Datasets for Building Segmentation for Raster Data
2.9.59. CVCMFFNet -> code for 2021 paper: Complex-Valued Convolutional and Multifeature Fusion Network for Building Semantic Segmentation of InSAR Images
2.9.60. STEB-UNet -> code for 2022 paper: A Swin Transformer-Based Encoding Booster Integrated in U-Shaped Network for Building Extraction
2.9.61. dfc2020_baseline -> Baseline solution for the IEEE GRSS Data Fusion Contest 2020. Predict land cover labels from Sentinel-1 and Sentinel-2 imagery. Code for 2020 paper: Weakly Supervised Semantic Segmentation of Satellite Images for Land Cover Mapping
2.9.62. Fusing multiple segmentation models based on different datasets into a single edge-deployable model -> roof, car & road segmentation
2.9.63. ground-truth-gan-segmentation -> use Pix2Pix to segment the footprint of a building. The dataset used is AIRS
2.9.64. UNICEF-Giga_Sudan -> Detecting school lots from satellite imagery in Southern Sudan using a UNET segmentation model
2.9.65. building_footprint_extraction -> The project retrieves satellite imagery from Google and performs building footprint extraction using a U-Net.
2.9.66. projectRegularization -> code for 2019 paper: Regularization of building boundaries in satellite images using adversarial and regularized losses
2.9.67. PolyWorldPretrainedNetwork -> code for 2021 paper: Polygonal Building Extraction with Graph Neural Networks in Satellite Images
2.9.68. dl_image_segmentation -> code for 2022 paper: Uncertainty-Aware Interpretable Deep Learning for Slum Mapping and Monitoring. Uses SHAP
2.9.69. UBC-dataset -> a dataset for building detection and classification from very high-resolution satellite imagery with the focus on object-level interpretation of individual buildings
2.9.70. GeoSeg -> code for 2022 paper: UNetFormer: A UNet-like transformer for efficient semantic segmentation of remote sensing urban scene imagery
2.9.71. BESNet -> code for 2022 paper: BES-Net: Boundary Enhancing Semantic Context Network for High-Resolution Image Semantic Segmentation. Applied to Vaihingen and Potsdam datasets
2.9.72. CVNet -> code for 2022 paper: CVNet: Contour Vibration Network for Building Extraction
2.9.73. CFENet -> code for 2022 paper: A Context Feature Enhancement Network for Building Extraction from High-Resolution Remote Sensing Imagery
2.9.74. HiSup -> code for 2022 paper: Accurate Polygonal Mapping of Buildings in Satellite Imagery
2.9.75. BuildingExtraction -> code for 2021 paper: Building Extraction from Remote Sensing Images with Sparse Token Transformers
2.9.76. coseg_building -> code for the 2022 paper: CrossGeoNet: A Framework for Building Footprint Generation of Label-Scarce Geographical Regions
2.9.77. AFM_building -> code for 2021 paper: Building Footprint Generation Through Convolutional Neural Networks With Attraction Field Representation
2.9.78. ramp-code -> code for the RAMP (Replicable AI for MicroPlanning) project, which enables building detection in low and middle income countries
2.9.79. Building-instance-segmentation -> code for 2022 paper: Multi-Modal Feature Fusion Network with Adaptive Center Point Detector for Building Instance Extraction
2.9.80. CGSANet -> code for the 2021 paper: CGSANet: A Contour-Guided and Local Structure-Aware Encoder–Decoder Network for Accurate Building Extraction From Very High-Resolution Remote Sensing Imagery
2.9.81. building-footprints-update -> code for 2022 paper: Learning Color Distributions from Bitemporal Remote Sensing Images to Update Existing Building Footprints
2.9.82. Istanbul_Dataset -> this repo contains weights of Unet++ model with SE-ResNeXt101 encoder trained with Istanbul, Inria and Massachusetts datasets seperately. Accompanies the paper: Comparative analysis of deep learning based building extraction methods with the new VHR Istanbul dataset
2.9.83. RAMP -> model and buildings dataset to support a wide variety of humanitarian use cases
2.9.84. Thesis_Semantic_Image_Segmentation_on_Satellite_Imagery_using_UNets -> This master thesis aims to perform semantic segmentation of buildings on satellite images from the SpaceNet challenge 1 dataset using the U-Net architecture
2.10. Segmentation - Solar panels
2.10.1. DeepSolar -> A Machine Learning Framework to Efficiently Construct a Solar Deployment Database in the United States. Dataset on kaggle, actually used a CNN for classification and segmentation is obtained by applying a threshold to the activation map. Original code is tf1 but tf2/kers and a pytorch implementation are available. Also checkout Visualizations and in-depth analysis .. of the factors that can explain the adoption of solar energy in .. Virginia and DeepSolar tracker: towards unsupervised assessment with open-source data of the accuracy of deep learning-based distributed PV mapping
2.10.2. hyperion_solar_net -> trained classificaton & segmentation models on RGB imagery from Google Maps. Provides app for viewing predictions, and has arxiv paper
2.10.3. 3D-PV-Locator -> Large-scale detection of rooftop-mounted photovoltaic systems in 3D
2.10.4. PV_Pipeline -> PyTorch models and pipeline developed for "DeepSolar for Germany"
2.10.5. solar-panels-detection -> using SegNet, Fast SCNN & ResNet
2.10.6. predict_pv_yield -> Using optical flow & machine learning to predict PV yield
2.10.7. Large-scale-solar-plant-monitoring -> code for the paper "Remote Sensing for Monitoring of Photovoltaic Power Plants in Brazil Using Deep Semantic Segmentation"
2.10.8. Panel-Segmentation -> Determine the presence of a solar array in the satellite image (boolean True/False), using a VGG16 classification model
2.10.9. Roofpedia -> an open registry of green roofs and solar roofs across the globe identified by Roofpedia through deep learning
2.10.10. Predicting the Solar Potential of Rooftops using Image Segmentation and Structured Data Medium article, using 20cm imagery & Unet
2.10.11. solar-pv-global-inventory -> code from the Nature paper of Kruitwagen et al, used to produce a global inventory of utility-scale solar photvoltaic generating stations
2.10.12. remote-sensing-solar-pv -> A repository for sharing progress on the automated detection of solar PV arrays in sentinel-2 remote sensing imagery
2.10.13. solar-panel-segmentation) -> Finding solar panels using USGS satellite imagery
2.10.14. solar_seg -> Solar segmentation of PV modules (sub elements of panels) using drone images and fast.ai
2.10.15. solar_plant_detection -> boundary extraction of Photovoltaic (PV) plants using Mask RCNN and Amir dataset
2.10.16. SolarDetection -> unet on satellite image from the USA and France
2.10.17. adopptrs -> Automatic Detection Of Photovoltaic Panels Through Remote Sensing using unet & pytorch
2.10.18. solar-panel-locator -> the number of solar panel pixels was only ~0.2% of the total pixels in the dataset, so solar panel data was upsampled to account for the class imbalance
2.10.19. projects-solar-panel-detection -> List of project to detect solar panels from aerial/satellite images
2.10.20. Satellite_ComputerVision -> UNET to detect solar arrays from Sentinel-2 data, using Google Earth Engine and Tensorflow. Also covers parking lot detection
2.10.21. photovoltaic-detection -> Detecting available rooftop area from satellite images to install photovoltaic panels
2.10.22. Solar_UNet -> U-Net models delineating solar arrays in Sentinel-2 imagery
2.11. Segmentation - Other manmade
2.11.1. Aarsh2001/ML_Challenge_NRSC -> Electrical Substation detection
2.11.2. electrical_substation_detection -> using UNet, Albumentations for image augmentation, and OpenCV for computer vision tasks
2.11.3. PLGAN-for-Power-Line-Segmentation -> code for 2022 paper: PLGAN: Generative Adversarial Networks for Power-Line Segmentation in Aerial Images
2.11.4. MCAN-OilSpillDetection -> Oil Spill Detection with A Multiscale Conditional Adversarial Network under Small Data Training, with paper. A multiscale conditional adversarial network (MCAN) trained with four oil spill observation images accurately detects oil spills in new images.
2.11.5. plastics -> Detecting and Monitoring Plastic Waste Aggregations in Sentinel-2 Imagery for globalplasticwatch.org
2.11.6. mining-detector -> detection of artisanal gold mines in Sentinel-2 satellite imagery for Amazon Mining Watch. Also covers clandestine airstrips
2.11.7. EG-UNet code for 2023 paper: Deep Feature Enhancement Method for Land Cover With Irregular and Sparse Spatial Distribution Features: A Case Study on Open-Pit Mining
2.12. Panoptic segmentation
2.12.1. Things and stuff or how remote sensing could benefit from panoptic segmentation
2.12.2. Panoptic Segmentation Meets Remote Sensing (paper)
2.12.3. pastis-benchmark
2.12.4. Panoptic-Generator -> This module converts GIS data into panoptic segmentation tiles
2.12.5. BSB-Aerial-Dataset -> an example on how to use Detectron2's Panoptic-FPN in the BSB Aerial Dataset
2.12.6. utae-paps -> PyTorch implementation of U-TAE and PaPs for satellite image time series panoptic segmentation
2.13. Segmentation - Miscellaneous
2.13.1. awesome-satellite-images-segmentation
2.13.2. Satellite Image Segmentation: a Workflow with U-Net is a decent intro article BEGINNER
2.13.3. mmsegmentation -> Semantic Segmentation Toolbox with support for many remote sensing datasets including LoveDA , Potsdam, Vaihingen & iSAID
2.13.4. segmentation_gym -> A neural gym for training deep learning models to carry out geoscientific image segmentation
2.13.5. How to create a DataBlock for Multispectral Satellite Image Semantic Segmentation using Fastai
2.13.6. Using a U-Net for image segmentation, blending predicted patches smoothly is a must to please the human eye -> python code to blend predicted patches smoothly. See Satellite-Image-Segmentation-with-Smooth-Blending
2.13.7. DCA -> code for 2022 paper: Deep Covariance Alignment for Domain Adaptive Remote Sensing Image Segmentation
2.13.8. SCAttNet -> Semantic Segmentation Network with Spatial and Channel Attention Mechanism
2.13.9. unetseg -> A set of classes and CLI tools for training a semantic segmentation model based on the U-Net architecture, using Tensorflow and Keras. This implementation is tuned specifically for satellite imagery and other geospatial raster data
2.13.10. Semantic Segmentation of Satellite Imagery using U-Net & fast.ai -> with repo
2.13.11. clusternet_segmentation -> Unsupervised Segmentation by applying K-Means clustering to the features generated by Neural Network
2.13.12. Collection of different Unet Variant -> demonstrates VggUnet, ResUnet, DenseUnet, Unet. AttUnet, MobileNetUnet, NestedUNet, R2AttUNet, R2UNet, SEUnet, scSEUnet, Unet_Xception_ResNetBlock, in keras
2.13.13. Efficient-Transformer -> code for 2021 paper: Efficient Transformer for Remote Sensing Image Segmentation
2.13.14. weakly_supervised -> code for the 2020 paper: Weakly Supervised Deep Learning for Segmentation of Remote Sensing Imagery
2.13.15. HRCNet-High-Resolution-Context-Extraction-Network -> code to 2021 paper: High-Resolution Context Extraction Network for Semantic Segmentation of Remote Sensing Images
2.13.16. Semantic segmentation of SAR images using a self supervised technique
2.13.17. satellite-segmentation-pytorch -> explores a wide variety of image augmentations to increase training dataset size
2.13.18. IEEE_TGRS_SpectralFormer -> code for 2021 paper: Spectralformer: Rethinking hyperspectral image classification with transformers
2.13.19. Unsupervised Segmentation of Hyperspectral Remote Sensing Images with Superpixels -> code for 2022 paper
2.13.20. Semantic-Segmentation-with-Sparse-Labels -> codes and data for learning from sparse annotations
2.13.21. SNDF -> code for 2020 paper: Superpixel-enhanced deep neural forest for remote sensing image semantic segmentation
2.13.22. Satellite-Image-Classification -> using random forest or support vector machines (SVM) and sklearn
2.13.23. dynamic-rs-segmentation -> code for 2019 paper: Dynamic Multi-Context Segmentation of Remote Sensing Images based on Convolutional Networks
2.13.24. Remote-sensing-image-semantic-segmentation-tf2 -> remote sensing image semantic segmentation repository based on tf.keras includes backbone networks such as resnet, densenet, mobilenet, and segmentation networks such as deeplabv3+, pspnet, panet, and refinenet
2.13.25. segmentation_models.pytorch -> Segmentation models with pretrained backbones, has been used in multiple winning solutions to remote sensing competitions
2.13.26. SSRN -> code for 2017 paper: Spectral-Spatial Residual Network for Hyperspectral Image Classification: A 3-D Deep Learning Framework
2.13.27. SO-DNN -> code for 2021 paper: Simplified object-based deep neural network for very high resolution remote sensing image classification
2.13.28. SANet -> code for 2019 paper: Scale-Aware Network for Semantic Segmentation of High-Resolution Aerial Images
2.13.29. aerial-segmentation -> code for 2017 paper: Learning Aerial Image Segmentation from Online Maps
2.13.30. IterativeSegmentation -> code for 2016 paper: Recurrent Neural Networks to Correct Satellite Image Classification Maps
2.13.31. Detectron2 FPN + PointRend Model for amazing Satellite Image Segmentation -> 15% increase in accuracy when compared to the U-Net model
2.13.32. HybridSN -> code for 2019 paper: HybridSN: Exploring 3D-2D CNN Feature Hierarchy for Hyperspectral Image Classification. Also a pytorch implementation here
2.13.33. TNNLS_2022_X-GPN -> code for 2022 paper: Semisupervised Cross-scale Graph Prototypical Network for Hyperspectral Image Classification
2.13.34. singleSceneSemSegTgrs2022 -> code for 2022 paper: Unsupervised Single-Scene Semantic Segmentation for Earth Observation
2.13.35. A-Fast-and-Compact-3-D-CNN-for-HSIC -> code for 2020 paper: A Fast and Compact 3-D CNN for Hyperspectral Image Classification
2.13.36. HSNRS -> code for 2017 paper: Hourglass-ShapeNetwork Based Semantic Segmentation for High Resolution Aerial Imagery
2.13.37. GiGCN -> code for 2022 paper: Graph-in-Graph Convolutional Network for Hyperspectral Image Classification
2.13.38. SSAN -> code for 2019 paper: Spectral-Spatial Attention Networks for Hyperspectral Image Classification
2.13.39. drone-images-semantic-segmentation -> Multiclass Semantic Segmentation of Aerial Drone Images Using Deep Learning
2.13.40. Satellite-Image-Segmentation-with-Smooth-Blending -> uses Smoothly-Blend-Image-Patches
2.13.41. BayesianUNet -> Pytorch Bayesian UNet model for segmentation and uncertainty prediction, applied to the Potsdam Dataset
2.13.42. RAANet -> code for 2022 paper: RAANet: A Residual ASPP with Attention Framework for Semantic Segmentation of High-Resolution Remote Sensing Images
2.13.43. wheelRuts_semanticSegmentation -> code for 2022 paper: Mapping wheel-ruts from timber harvesting operations using deep learning techniques in drone imagery
2.13.44. LWN-for-UAVRSI -> Light-Weight Semantic Segmentation Network for UAV Remote Sensing Images, applied to Vaihingen, UAVid and UDD6 datasets
2.13.45. hypernet -> library which implements; accurate hyperspectral image (HSI) segmentation and analysis using deep neural networks, optimization of deep neural network architectures for hyperspectral data segmentation, hyperspectral data augmentation, validation of existent and emerging HSI segmentation algorithms, simulation of multispectral data using HSI
2.13.46. ST-UNet -> code for 2022 paper: Swin Transformer Embedding UNet for Remote Sensing Image Semantic Segmentation
2.13.47. EDFT -> code for 2022 paper: Efficient Depth Fusion Transformer for Aerial Image Semantic Segmentation
2.13.48. WiCoNet -> code for 2022 paper: Looking Outside the Window: Wide-Context Transformer for the Semantic Segmentation of High-Resolution Remote Sensing Images
2.13.49. CRGNet -> code for 2022 paper: Consistency-Regularized Region-Growing Network for Semantic Segmentation of Urban Scenes with Point-Level Annotations
2.13.50. SA-UNet -> code for 2022 paper: Improved U-Net Remote Sensing Classification Algorithm Fusing Attention and Multiscale Features
2.13.51. MANet -> code for 2020 paper: Multi-Attention-Network for Semantic Segmentation of Fine Resolution Remote Sensing Images
2.13.52. BANet -> code for 2021 paper: Transformer Meets Convolution: A Bilateral Awareness Network for Semantic Segmentation of Very Fine Resolution Urban Scene Images
2.13.53. MACU-Net -> code for 2022 paper: MACU-Net for Semantic Segmentation of Fine-Resolution Remotely Sensed Images
2.13.54. DNAS -> code for 2022 paper: DNAS: Decoupling Neural Architecture Search for High-Resolution Remote Sensing Image Semantic Segmentation
2.13.55. A2-FPN -> code for 2021 paper: A2-FPN for Semantic Segmentation of Fine-Resolution Remotely Sensed Images
2.13.56. MAResU-Net -> code for 2020 paper: Multi-stage Attention ResU-Net for Semantic Segmentation of Fine-Resolution Remote Sensing Images
2.13.57. ml_segmentation -> semantic segmentation of buildings using Random Forest, Support Vector Machine (SVM) & Gradient Boosting Classifier (GBC)
2.13.58. RSEN -> code for 2021 paper: Robust Self-Ensembling Network for Hyperspectral Image Classification
2.13.59. MSNet -> code for 2022 paper: MSNet: multispectral semantic segmentation network for remote sensing images
2.13.60. k-textures -> code (R) for 2022 paper: K-textures, a self-supervised hard clustering deep learning algorithm for satellite image segmentation
2.13.61. Swin-Transformer-Semantic-Segmentation -> code for 2021 paper: Satellite Image Semantic Segmentation
2.13.62. UDA_for_RS -> code for 2022 paper: Unsupervised Domain Adaptation for Remote Sensing Semantic Segmentation with Transformer
2.13.63. A-3D-CNN-AM-DSC-model-for-hyperspectral-image-classification -> code for 2022 paper: Attention Mechanism and Depthwise Separable Convolution Aided 3DCNN for Hyperspectral Remote Sensing Image Classification
2.13.64. contrastive-distillation -> code for paper: A Contrastive Distillation Approach for Incremental Semantic Segmentation in Aerial Images
2.13.65. SegForestNet -> code for 2023 paper: SegForestNet: Spatial-Partitioning-Based Aerial Image Segmentation
2.13.66. MFVNet -> code for 2023 paper: MFVNet: Deep Adaptive Fusion Network with Multiple Field-of-Views for Remote Sensing Image Semantic Segmentation
2.13.67. Wildebeest-UNet -> detecting wildebeest and zebras in Serengeti-Mara ecosystem from very-high-resolution satellite imagery
2.13.68. segment-anything-eo -> Earth observation tools for Meta AI Segment Anything (SAM - Segment Anything Model)
2.13.69. HR-Image-classification_SDF2N -> code for 2023 paper: A Shallow-to-Deep Feature Fusion Network for VHR Remote Sensing Image Classification
3. Instance segmentation
In instance segmentation, each individual 'instance' of a segmented area is given a unique lable. For detection of very small objects this may a good approach, but it can struggle seperating individual objects that are closely spaced.
3.1. Mask_RCNN generates bounding boxes and segmentation masks for each instance of an object in the image. It is very commonly used for instance segmentation & object detection
3.2. Instance segmentation of center pivot irrigation system in Brazil using free Landsat images, mask R-CNN & Keras BEGINNER
3.3. Building-Detection-MaskRCNN -> Building detection from the SpaceNet dataset by using Mask RCNN BEGINNER
3.4. Oil tank instance segmentation with Mask R-CNN with accompanying article using Keras & Airbus Oil Storage Detection Dataset on Kaggle
3.5. Mask_RCNN-for-Caravans -> detect caravan footprints from OS imagery
3.6. parking_bays_detectron2 -> Detecting parking bays with satellite imagery. Used Detectron2 and synthetic data with Unreal, superior performance to using Mask RCNN
3.7. Locate buildings with a dark roof that feed heat island phenomenon using Mask RCNN -> with repo, used INRIA dataset & labelme for annotation
3.8. Circle_Finder -> Circular Shapes Detection in Satellite Imagery, 2nd place solution to the Circle Finder Challenge
3.9. Lawn_maskRCNN -> Detecting lawns from satellite images of properties in the Cedar Rapids area using Mask-R-CNN
3.10. CropMask_RCNN -> Segmenting center pivot agriculture to monitor crop water use in drylands with Mask R-CNN and Landsat satellite imagery
3.11. Mask RCNN for Spacenet Off Nadir Building Detection
3.12. CATNet -> code for 2021 paper: Learning to Aggregate Multi-Scale Context for Instance Segmentation in Remote Sensing Images
3.13. Object-Detection-on-Satellite-Images-using-Mask-R-CNN -> detect ships
3.14. FactSeg -> Foreground Activation Driven Small Object Semantic Segmentation in Large-Scale Remote Sensing Imagery (TGRS), also see FarSeg and FreeNet, implementations of research paper
3.15. aqua_python -> detecting aquaculture farms using Mask R-CNN
3.16. RSPrompter -> code for 2023 paper: Learning to Prompt for Remote Sensing Instance Segmentation based on Visual Foundation Model
4. Object detection
Image showing the suitability of rotated bounding boxes in remote sensing.
Object detection in remote sensing involves locating and surrounding objects of interest with bounding boxes. Due to the large size of remote sensing images and the fact that objects may only comprise a few pixels, object detection can be challenging in this context. The imbalance between the area of the objects to be detected and the background, combined with the potential for objects to be easily confused with random features in the background, further complicates the task. Object detection generally performs better on larger objects, but becomes increasingly difficult as the objects become smaller and more densely packed. The accuracy of object detection models can also degrade rapidly as image resolution decreases, which is why it is common to use high resolution imagery, such as 30cm RGB, for object detection in remote sensing. A unique characteristic of aerial images is that objects can be oriented in any direction. To effectively extract measurements of the length and width of an object, it can be crucial to use rotated bounding boxes that align with the orientation of the object. This approach enables more accurate and meaningful analysis of the objects within the image. Image source
4.1. Object tracking in videos
4.1.1. Object Tracking in Satellite Videos Based on a Multi-Frame Optical Flow Tracker arxiv paper
4.1.2. CFME -> Object Tracking in Satellite Videos by Improved Correlation Filters With Motion Estimations
4.1.3. TGraM -> code and dataset for 2022 paper: Multi-Object Tracking in Satellite Videos with Graph-Based Multi-Task Modeling
4.1.4. satellite_video_mod_groundtruth -> groundtruth on satellite video for evaluating moving object detection algorithm
4.1.5. Moving-object-detection-DSFNet -> code for 2021 paper: DSFNet: Dynamic and Static Fusion Network for Moving Object Detection in Satellite Videos
4.1.6. HiFT -> code for 2021 paper: HiFT: Hierarchical Feature Transformer for Aerial Tracking
4.1.7. TCTrack -> code for 2022 paper: TCTrack: Temporal Contexts for Aerial Tracking
4.2. Object detection with rotated bounding boxes
Orinted bounding boxes (OBB) are polygons representing rotated rectangles. For datasets checkout DOTA & HRSC2016
4.2.1. mmrotate -> Rotated Object Detection Benchmark, with pretrained models and function for inferencing on very large images
4.2.2. OBBDetection -> an oriented object detection library, which is based on MMdetection
4.2.3. rotate-yolov3 -> Rotation object detection implemented with yolov3. Also see yolov3-polygon
4.2.4. DRBox -> for detection tasks where the objects are orientated arbitrarily, e.g. vehicles, ships and airplanes
4.2.5. s2anet -> Official code of the paper 'Align Deep Features for Oriented Object Detection'
4.2.6. CFC-Net -> Official implementation of "CFC-Net: A Critical Feature Capturing Network for Arbitrary-Oriented Object Detection in Remote Sensing Images"
4.2.7. ReDet -> Official code of the paper "ReDet: A Rotation-equivariant Detector for Aerial Object Detection"
4.2.8. BBAVectors-Oriented-Object-Detection -> Oriented Object Detection in Aerial Images with Box Boundary-Aware Vectors
4.2.9. CSL_RetinaNet_Tensorflow -> Code for ECCV 2020 paper: Arbitrary-Oriented Object Detection with Circular Smooth Label
4.2.10. r3det-on-mmdetection -> R3Det: Refined Single-Stage Detector with Feature Refinement for Rotating Object
4.2.11. R-DFPN_FPN_Tensorflow -> Rotation Dense Feature Pyramid Networks (Tensorflow)
4.2.12. R2CNN_Faster-RCNN_Tensorflow -> Rotational region detection based on Faster-RCNN
4.2.13. Rotated-RetinaNet -> implemented in pytorch, it supports the following datasets: DOTA, HRSC2016, ICDAR2013, ICDAR2015, UCAS-AOD, NWPU VHR-10, VOC2007
4.2.14. OBBDet_Swin -> The sixth place winning solution in 2021 Gaofen Challenge
4.2.15. CG-Net -> Learning Calibrated-Guidance for Object Detection in Aerial Images. With paper
4.2.16. OrientedRepPoints_DOTA -> Oriented RepPoints + Swin Transformer/ReResNet
4.2.17. yolov5_obb -> yolov5 + Oriented Object Detection
4.2.18. How to Train YOLOv5 OBB -> YOLOv5 OBB tutorial and YOLOv5 OBB noteboook
4.2.19. OHDet_Tensorflow -> can be applied to rotation detection and object heading detection
4.2.20. Seodore -> framework maintaining recent updates of mmdetection
4.2.21. Rotation-RetinaNet-PyTorch -> oriented detector Rotation-RetinaNet implementation on Optical and SAR ship dataset
4.2.22. AIDet -> an open source object detection in aerial image toolbox based on MMDetection
4.2.23. rotation-yolov5 -> rotation detection based on yolov5
4.2.24. ShipDetection -> Ship Detection in HR Optical Remote Sensing Images via Rotated Bounding Box, based on Faster R-CNN and ORN, uses caffe
4.2.25. SLRDet -> project based on mmdetection to reimplement RRPN and use the model Faster R-CNN OBB
4.2.26. AxisLearning -> code for 2020 paper: Axis Learning for Orientated Objects Detection in Aerial Images
4.2.27. Detection_and_Recognition_in_Remote_Sensing_Image -> This work uses PaNet to realize Detection and Recognition in Remote Sensing Image by MXNet
4.2.28. DrBox-v2-tensorflow -> tensorflow implementation of DrBox-v2 which is an improved detector with rotatable boxes for target detection in remote sensing images
4.2.29. Rotation-EfficientDet-D0 -> A PyTorch Implementation Rotation Detector based EfficientDet Detector, applied to custom rotation vehicle datasets
4.2.30. DODet -> Dual alignment for oriented object detection, uses DOTA dataset. With paper
4.2.31. GF-CSL -> code for 2022 paper: Gaussian Focal Loss: Learning Distribution Polarized Angle Prediction for Rotated Object Detection in Aerial Images
4.2.32. simplified_rbox_cnn -> code for 2018 paper: RBox-CNN: rotated bounding box based CNN for ship detection in remote sensing image. Uses Tensorflow object detection API
4.2.33. Polar-Encodings -> code for 2021 [paper](Learning Polar Encodings for Arbitrary-Oriented Ship Detection in SAR Images)
4.2.34. R-CenterNet -> detector for rotated-object based on CenterNet
4.2.35. piou -> Orientated Object Detection; IoU Loss, applied to DOTA dataset
4.2.36. DAFNe -> code for 2021 paper: DAFNe: A One-Stage Anchor-Free Approach for Oriented Object Detection
4.2.37. AProNet -> code for 2021 paper: AProNet: Detecting objects with precise orientation from aerial images. Applied to datasets DOTA and HRSC2016
4.2.38. UCAS-AOD-benchmark -> A benchmark of UCAS-AOD dataset
4.2.39. RotateObjectDetection -> based on Ultralytics/yolov5, with adjustments to enable rotate prediction boxes. Also see PolygonObjectDetection
4.2.40. AD-Toolbox -> Aerial Detection Toolbox based on MMDetection and MMRotate, with support for more datasets
4.2.41. GGHL -> code for 2022 paper: A General Gaussian Heatmap Label Assignment for Arbitrary-Oriented Object Detection
4.2.42. NPMMR-Det -> code for 2021 paper: A Novel Nonlocal-Aware Pyramid and Multiscale Multitask Refinement Detector for Object Detection in Remote Sensing Images
4.2.43. AOPG -> code for 2022 paper: Anchor-Free Oriented Proposal Generator for Object Detection
4.2.44. SE2-Det -> code for 2022 paper: Semantic-Edge-Supervised Single-Stage Detector for Oriented Object Detection in Remote Sensing Imagery
4.2.45. OrientedRepPoints -> code for 2021 paper: Oriented RepPoints for Aerial Object Detection
4.2.46. TS-Conv -> code for 2022 paper: Task-wise Sampling Convolutions for Arbitrary-Oriented Object Detection in Aerial Images
4.2.47. FCOSR -> A Simple Anchor-free Rotated Detector for Aerial Object Detection. This implement is modified from mmdetection. See also TensorRT_Inference
4.2.48. OBB_Detection -> Finalist's solution in the track of Oriented Object Detection in Remote Sensing Images, 2022 Guangdong-Hong Kong-Macao Greater Bay Area International Algorithm Competition
4.2.49. sam-mmrotate -> SAM (Segment Anything Model) for generating rotated bounding boxes with MMRotate, which is a comparison method of H2RBox-v2
4.2.50. mmrotate-dcfl -> code for 2023 paper: Dynamic Coarse-to-Fine Learning for Oriented Tiny Object Detection
4.2.51. h2rbox-mmrotate -> code for 2022 paper: H2RBox: Horizontal Box Annotation is All You Need for Oriented Object Detection
4.3. Object detection enhanced by super resolution
4.3.1. Super-Resolution and Object Detection -> Super-resolution is a relatively inexpensive enhancement that can improve object detection performance
4.3.2. EESRGAN -> Small-Object Detection in Remote Sensing Images with End-to-End Edge-Enhanced GAN and Object Detector Network
4.3.3. Mid-Low Resolution Remote Sensing Ship Detection Using Super-Resolved Feature Representation
4.3.4. EESRGAN -> code for 2020 paper: Small-Object Detection in Remote Sensing Images with End-to-End Edge-Enhanced GAN and Object Detector Network. Applied to COWC & OGST datasets
4.3.5. FBNet -> code for 2022 paper: Feature Balance for Fine-Grained Object Classification in Aerial Images
4.3.6. SuperYOLO -> code for 2022 paper: SuperYOLO: Super Resolution Assisted Object Detection in Multimodal Remote Sensing Imagery
4.4. Salient object detection
Detecting the most noticeable or important object in a scene
4.4.1. ACCoNet -> code for 2022 paper: Adjacent Context Coordination Network for Salient Object Detection in Optical Remote Sensing Images
4.4.2. MCCNet -> Multi-Content Complementation Network for Salient Object Detection in Optical Remote Sensing Images
4.4.3. CorrNet -> Lightweight Salient Object Detection in Optical Remote Sensing Images via Feature Correlation. With paper
4.4.4. Reading list for deep learning based Salient Object Detection in Optical Remote Sensing Images
4.4.5. ORSSD-dataset -> salient object detection dataset
4.4.6. EORSSD-dataset -> Extended Optical Remote Sensing Saliency Detection (EORSSD) Dataset
4.4.7. DAFNet_TIP20 -> code for 2020 paper: Dense Attention Fluid Network for Salient Object Detection in Optical Remote Sensing Images
4.4.8. EMFINet -> code for 2021 paper: Edge-Aware Multiscale Feature Integration Network for Salient Object Detection in Optical Remote Sensing Images
4.4.9. ERPNet -> code for 2022 paper: Edge-guided Recurrent Positioning Network for Salient Object Detection in Optical Remote Sensing Images
4.4.10. FSMINet -> code for 2022 paper: Fully Squeezed Multi-Scale Inference Network for Fast and Accurate Saliency Detection in Optical Remote Sensing Images
4.4.11. AGNet -> code for 2022 paper: AGNet: Attention Guided Network for Salient Object Detection in Optical Remote Sensing Images
4.4.12. MSCNet -> code for 2022 paper: A lightweight multi-scale context network for salient object detection in optical remote sensing images
4.4.13. GPnet -> code for 2022 paper: Global Perception Network for Salient Object Detection in Remote Sensing Images
4.4.14. SeaNet -> code for 2023 paper: Lightweight Salient Object Detection in Optical Remote Sensing Images via Semantic Matching and Edge Alignment
4.5. Object detection - Buildings, rooftops & solar panels
4.5.1. satellite_image_tinhouse_detector -> Detection of tin houses from satellite/aerial images using the Tensorflow Object Detection API BEGINNER
4.5.2. Machine Learning For Rooftop Detection and Solar Panel Installment discusses tiling large images and generating annotations from OSM data. Features of the roofs were calculated using a combination of contour detection and classification. Follow up article using semantic segmentation
4.5.3. Building Extraction with YOLT2 and SpaceNet Data
4.5.4. XBD-hurricanes -> Models for building (and building damage) detection in high-resolution (<1m) satellite and aerial imagery using a modified RetinaNet model
4.5.5. Detecting solar panels from satellite imagery using segmentation
4.5.6. ssd-spacenet -> Detect buildings in the Spacenet dataset using Single Shot MultiBox Detector (SSD)
4.5.7. 3DBuildingInfoMap -> simultaneous extraction of building height and footprint from Sentinel imagery using ResNet
4.5.8. Solar Panel Detection -> using Faster R-CNN & Tensorflow object detection API. With repo
4.5.9. DeepSolaris -> a EuroStat project to detect solar panels in aerial images, further material here
4.5.10. ML_ObjectDetection_CAFO -> Detect Concentrated Animal Feeding Operations (CAFO) in Satellite Imagery
4.5.11. Multi-level-Building-Detection-Framework -> code for 2018 paper: Multilevel Building Detection Framework in Remote Sensing Images Based on Convolutional Neural Networks
4.5.12. Automatic Damage Annotation on Post-Hurricane Satellite Imagery -> detect damaged buildings using tensorflow object detection API. With repos here and here
4.5.13. mappingchallenge -> YOLOv5 applied to the AICrowd Mapping Challenge dataset
4.6. Object detection - Ships & boats
4.6.1. kaggle-ships-in-Google-Earth-yolov5 -> Applying YOLOv5 to Kaggle Ships in Google Earth dataset BEGINNER
4.6.2. How hard is it for an AI to detect ships on satellite images?
4.6.3. Object Detection in Satellite Imagery, a Low Overhead Approach
4.6.4. Detecting Ships in Satellite Imagery using the Planet dataset and Keras
4.6.5. Planet use non DL felzenszwalb algorithm to detect ships
4.6.6. Ship detection using k-means clustering & CNN classifier on patches
4.6.7. SARfish -> Ship detection in Sentinel 1 Synthetic Aperture Radar (SAR) imagery
4.6.8. Arbitrary-Oriented Ship Detection through Center-Head Point Extraction -> arxiv paper. Keypoint estimation is performed to find the center of ships. Then, the size and head point of the ships are regressed. Repo ASD
4.6.9. ship_detection -> using an interesting combination of CNN classifier, Class Activation Mapping (CAM) & UNET segmentation. Accompanying three part blog post
4.6.10. Building a complete Ship detection algorithm using YOLOv3 and Planet satellite images -> covers finding and annotating data (using LabelMe), preprocessing large images into chips, and training Yolov3. Repo
4.6.11. Ship-detection-in-satellite-images -> experiments with UNET, YOLO, Mask R-CNN, SSD, Faster R-CNN, RETINA-NET
4.6.12. Ship-Detection-from-Satellite-Images-using-YOLOV4 -> uses Kaggle Airbus Ship Detection dataset
4.6.13. kaggle-airbus-ship-detection-challenge -> using oriented SSD
4.6.14. shipsnet-detector -> Detect container ships in Planet imagery using machine learning
4.6.15. Classifying Ships in Satellite Imagery with Neural Networks -> applied to the Kaggle Ships in Satellite Imagery dataset
4.6.16. Mask R-CNN for Ship Detection & Segmentation blog post with repo
4.6.17. contrastive_SSL_ship_detection -> Contrastive self supervised learning for ship detection in Sentinel 2 images
4.6.18. Boat detection with multi-region-growing method in satellite images
4.6.19. small-boat-detector -> Trained yolo v3 model weights and configuration file to detect small boats in satellite imagery
4.6.20. Satellite-Imagery-Datasets-Containing-Ships -> A list of optical and radar satellite datasets for ship detection, classification, semantic segmentation and instance segmentation tasks
4.6.21. Ship-Classification-in-Satellite-Images -> Convolutional neural network model for ship classification in satellite images
4.6.22. Ship-Detection -> CNN approach for ship detection in the ocean using a satellite image
4.6.23. vesselTracker -> Project based on reduced model of Yolov5 architecture using Pytorch. Custom dataset based on SAR imagery provided by Sentinel-1 through Earth Engine API
4.6.24. marine-debris-ml-model -> Marine Debris Detection using tensorflow object detection API
4.6.25. SDGH-Net -> code for 2021 paper: SDGH-Net: Ship Detection in Optical Remote Sensing Images Based on Gaussian Heatmap Regression
4.6.26. LR-TSDet -> code for 2021 paper: LR-TSDet: Towards Tiny Ship Detection in Low-Resolution Remote Sensing Images
4.6.27. FGSCR-42 -> A public Dataset for Fine-Grained Ship Classification in Remote sensing images
4.6.28. ShipDetection -> Ship Detection in HR Optical Remote Sensing Images via Rotated Bounding Box, based on Faster R-CNN and ORN, uses caffe
4.6.29. WakeNet -> A CNN-based optical image ship wake detector, code for 2021 paper: Rethinking Automatic Ship Wake Detection: State-of-the-Art CNN-based Wake Detection via Optical Images
4.6.30. Histogram of Oriented Gradients (HOG) Boat Heading Classification -> Medium article
4.6.31. Object Detection in Satellite Imagery, a Low Overhead Approach -> Medium article which demonstrates how to combine Canny edge detector pre-filters with HOG feature descriptors, random forest classifiers, and sliding windows to perform ship detection
4.6.32. simplified_rbox_cnn -> code for 2018 paper: RBox-CNN: rotated bounding box based CNN for ship detection in remote sensing image. Uses Tensorflow object detection API
4.6.33. Ship-Detection-based-on-YOLOv3-and-KV260 -> entry project of the Xilinx Adaptive Computing Challenge 2021. It uses YOLOv3 for ship target detection in optical remote sensing images, and deploys DPU on the KV260 platform to achieve hardware acceleration
4.6.34. LEVIR-Ship -> a dataset for tiny ship detection under medium-resolution remote sensing images
4.6.35. Push-and-Pull-Network -> code for 2022 paper: Contrastive Learning for Fine-grained Ship Classification in Remote Sensing Images
4.6.36. DRENet -> code for 2022 [paper])(https://ieeexplore.ieee.org/abstract/document/9791363): A Degraded Reconstruction Enhancement-Based Method for Tiny Ship Detection in Remote Sensing Images With a New Large-Scale Dataset
4.6.37. xView3-The-First-Place-Solution - A winning solution for xView 3 challenge (Vessel detection, classification and length estimation on Sentinetl-1 images). Contains trained models, inference pipeline and training code & configs to reproduce the results.
4.7. Object detection - Cars, vehicles & trains
4.7.1. Detection of parkinglots and driveways with retinanet BEGINNER
4.7.2. pytorch-vedai -> object detection on the VEDAI dataset: Vehicle Detection in Aerial Imagery BEGINNER
4.7.3. Truck Detection with Sentinel-2 during COVID-19 crisis -> moving objects in Sentinel-2 data causes a specific reflectance relationship in the RGB, which looks like a rainbow, and serves as a marker for trucks. Improve accuracy by only analysing roads. Not using object detection but relevant. Also see S2TD
4.7.4. cowc_car_counting -> car counting on the Cars Overhead With Context (COWC) dataset. Not sctictly object detection but a CNN to predict the car count in a tile
4.7.5. CarCounting -> using Yolov3 & COWC dataset
4.7.6. Traffic density estimation as a regression problem instead of object detection -> inspired by this paper
4.7.7. Applying Computer Vision to Railcar Detection -> useful insights into counting railcars (i.e. train carriages) using Mask-RCNN with rotated bounding boxes output
4.7.8. Leveraging Deep Learning for Vehicle Detection And Classification
4.7.9. Rotation-EfficientDet-D0 -> PyTorch implementation of Rotated EfficientDet, applied to a custom rotation vehicle dataset (car counting)
4.7.10. RSVC2021-Dataset -> A dataset for Vehicle Counting in Remote Sensing images, created from the DOTA & ITCVD
4.7.11. Car Localization and Counting with Overhead Imagery, an Interactive Exploration -> Medium article by Adam Van Etten
4.7.12. Vehicle-Counting-in-Very-Low-Resolution-Aerial-Images -> code for 2022 paper: Vehicle Counting in Very Low-Resolution Aerial Images via Cross-Resolution Spatial Consistency and Intraresolution Time Continuity
4.7.13. Vehicle Detection blog post by Grant Pearse: detecting vehicles across New Zealand without collecting local training data
4.8. Object detection - Planes & aircraft
4.8.1. Faster RCNN to detect airplanes BEGINNER
4.8.2. yoltv4 includes examples on the RarePlanes dataset
4.8.3. aircraft-detection -> experiments to test the performance of a Gaussian process (GP) classifier with various kernels on the UC Merced land use land cover (LULC) dataset
4.8.4. Using Detectron2 to segment aircraft from satellite imagery -> pytorch and Rare Planes
4.8.5. aircraft-detection-from-satellite-images-yolov3 -> trained on kaggle cgi-planes-in-satellite-imagery-w-bboxes dataset
4.8.6. HRPlanesv2-Data-Set -> YOLOv4 and YOLOv5 weights trained on the HRPlanesv2 dataset
4.8.7. Deep-Learning-for-Aircraft-Recognition -> A CNN model trained to classify and identify various military aircraft through satellite imagery
4.8.8. FRCNN-for-Aircraft-Detection -> faster-rcnn & keras
4.8.9. ergo-planes-detector -> An ergo based project that relies on a convolutional neural network to detect airplanes from satellite imagery, uses the PlanesNet dataset
4.8.10. pytorch-remote-sensing -> Aircraft detection using the 'Airbus Aircraft Detection' dataset and Faster-RCNN with ResNet-50 backbone using pytorch
4.8.11. FasterRCNN_ObjectDetection -> faster RCNN model for aircraft detection and localisation in satellite images and creating a webpage with live server for public usage
4.8.12. HRPlanes -> weights of YOLOv4 and Faster R-CNN networks trained with HRPlanes dataset
4.8.13. aerial-detection -> uses Yolov5 & Icevision
4.8.14. How to choose a deep learning architecture to detect aircrafts in satellite imagery?
4.8.15. rareplanes-yolov5 -> using YOLOv5 and the RarePlanes dataset to detect and classify sub-characteristics of aircraft, with article
4.8.16. OnlyPlanes -> dataset and pretrained models for the paper: OnlyPlanes - Incrementally Tuning Synthetic Training Datasets for Satellite Object Detection
4.8.17. Understanding the RarePlanes Dataset and Building an Aircraft Detection Model -> blog post
4.9. Object detection - Infrastructure & utilities
4.9.1. wind-turbine-detector -> Wind Turbine Object Detection from Aerial Imagery Using TensorFlow Object Detection API
4.9.2. Water Tanks and Swimming Pools Detection -> uses Faster R-CNN
4.9.3. PCAN -> Part-Based Context Attention Network for Thermal Power Plant Detection in Remote Sensing Imagery, with dataset
4.10. Object detection - Oil storage tank detection
Oil is stored in tanks at many points between extraction and sale, and the volume of oil in storage is an important economic indicator.
4.10.1. A Beginner’s Guide To Calculating Oil Storage Tank Occupancy With Help Of Satellite Imagery
4.10.2. Oil Storage Tank’s Volume Occupancy On Satellite Imagery Using YoloV3 with repo
4.10.3. Oil-Tank-Volume-Estimation -> combines object detection and classical computer vision
4.10.4. Oil tank instance segmentation with Mask R-CNN with accompanying article using Keras & Airbus Oil Storage Detection Dataset on Kaggle
4.10.5. https://www.kaggle.com/towardsentropy/oil-storage-tanks -> large kaggle dataset, note however that approx 85% of images contain no tanks
4.10.6. https://www.kaggle.com/airbusgeo/airbus-oil-storage-detection-dataset -> smaller kaggle dataset
4.10.7. ognet -> a Global Oil and Gas Infrastructure Database using Deep Learning on Remotely Sensed Imagery
4.10.8. RSOD-Dataset -> dataset for object detection in PASCAL VOC format. Aircraft, playgrounds, overpasses & oiltanks. Used in the 2022 paper: Improved YOLOv5 network method for remote sensing image-based ground objects recognition
4.10.9. oil_storage-detector -> using yolov5 and the Airbus Oil Storage Detection dataset
4.10.10. oil_well_detector -> detect oil wells in the Bakken oil field based on satellite imagery
4.10.11. OGST -> Oil and Gas Tank Dataset
4.10.12. AContrarioTankDetection -> code for 2020 paper: Oil Tank Detection in Satellite Images via a Contrario Clustering
4.10.13. SubpixelCircleDetection -> code for 2020 paper: CIRCULAR-SHAPED OBJECT DETECTION IN LOW RESOLUTION SATELLITE IMAGES
4.10.14. Oil Storage Detection on Airbus Imagery with YOLOX -> uses the Kaggle Airbus Oil Storage Detection dataset
4.11. Object detection - Animals
A variety of techniques can be used to count animals, including object detection and instance segmentation. For convenience they are all listed here:
4.11.1. cownter_strike -> counting cows, located with point-annotations, two models: CSRNet (a density-based method) & LCFCN (a detection-based method)
4.11.2. elephant_detection -> Using Keras-Retinanet to detect elephants from aerial images
4.11.3. CNN-Mosquito-Detection -> determining the locations of potentially dangerous breeding grounds, compared YOLOv4, YOLOR & YOLOv5
4.11.4. Borowicz_etal_Spacewhale -> locate whales using ResNet
4.11.5. walrus-detection-and-count -> uses Mask R-CNN instance segmentation
4.11.6. MarineMammalsDetection -> Weakly Supervised Detection of Marine Animals in High Resolution Aerial Images
4.11.7. Audubon_F21 -> code for 2022 paper: Deep object detection for waterbird monitoring using aerial imagery
4.12. Object detection - Miscellaneous
4.12.1. Object detection on Satellite Imagery using RetinaNet -> using the Kaggle Swimming Pool and Car Detection dataset BEGINNER
4.12.2. Tackling the Small Object Problem in Object Detection BEGINNER
4.12.3. Object Detection and Image Segmentation with Deep Learning on Earth Observation Data: A Review
4.12.4. awesome-aerial-object-detection bu murari023, another by visionxiang and awesome-tiny-object-detection list many relevant papers
4.12.5. Object Detection Accuracy as a Function of Image Resolution -> Medium article using COWC dataset, performance rapidly degrades below 30cm imagery
4.12.6. Satellite Imagery Multiscale Rapid Detection with Windowed Networks (SIMRDWN) -> combines some of the leading object detection algorithms into a unified framework designed to detect objects both large and small in overhead imagery. Train models and test on arbitrary image sizes with YOLO (versions 2 and 3), Faster R-CNN, SSD, or R-FCN.
4.12.7. YOLTv4 -> YOLTv4 is designed to detect objects in aerial or satellite imagery in arbitrarily large images that far exceed the ~600×600 pixel size typically ingested by deep learning object detection frameworks. Read Announcing YOLTv4: Improved Satellite Imagery Object Detection
4.12.8. Tensorflow Benchmarks for Object Detection in Aerial Images -> tensorflow-based codebase created to build benchmarks for object detection in aerial images
4.12.9. Pytorch Benchmarks for Object Detection in Aerial Images -> pytorch-based codebase created to build benchmarks for object detection in aerial images using mmdetection
4.12.10. ASPDNet -> Counting dense objects in remote sensing images, arxiv paper
4.12.11. xview-yolov3 -> xView 2018 Object Detection Challenge: YOLOv3 Training and Inference
4.12.12. Faster RCNN for xView satellite data challenge
4.12.13. How to detect small objects in (very) large images -> A practical guide to using Slicing-Aided Hyper Inference (SAHI) for performing inference on the DOTAv1.0 object detection dataset using the mmdetection framework
4.12.14. Object Detection Satellite Imagery Multi-vehicles Dataset (SIMD) -> RetinaNet,Yolov3 and Faster RCNN for multi object detection on satellite images dataset
4.12.15. SNIPER/AutoFocus -> an efficient multi-scale object detection training/inference algorithm
4.12.16. marine_debris_ML -> Marine debris detection, uses 3-meter imagery product called Planetscope with bands in the red, green, blue, and near-infrared. Uses Tensorflow Object Detection API with pre-trained resnet 101
4.12.17. pool-detection-from-aerial-imagery -> Use Icevision and Detectron2 to detect swimming pools from aerial imagery
4.12.18. Electric-Pylon-Detection-in-RSI -> a dataset which contains 1500 remote sensing images of electric pylons used to train ten deep learning models
4.12.19. Synthesizing Robustness YOLTv4 Results Part 2: Dataset Size Requirements and Geographic Insights -> quantify how much harder rare objects are to localize
4.12.20. IS-Count -> IS-Count is a sampling-based and learnable method for estimating the total object count in a region.
4.12.21. Object Detection On Aerial Imagery Using RetinaNet
4.12.22. Clustered-Object-Detection-in-Aerial-Image
4.12.23. yolov5s_for_satellite_imagery -> yolov5s applied to the DOTA dataset
4.12.24. RetinaNet-PyTorch -> RetinaNet implementation on remote sensing ship dataset (SSDD)
4.12.25. Detecting-Cyclone-Centers-Custom-YOLOv3 -> tropical cyclones (TCs) are intense warm-corded cyclonic vortices, developed from low-pressure systems over the tropical oceans and driven by complex air-sea interaction
4.12.26. Object-Detection-YoloV3-RetinaNet-FasterRCNN -> trained on a private datset
4.12.27. Google-earth-Object-Recognition -> Code for training and evaluating on Dior Dataset (Google Earth Images) using RetinaNet and YOLOV5
4.12.28. HIECTOR: Hierarchical object detector at scale -> HIECTOR facilitates multiple satellite data collections of increasingly detailed spatial resolution for a cost-efficient and accurate object detection over large areas. Code on Github
4.12.29. Detection of Multiclass Objects in Optical Remote Sensing Images -> code for 2018 paper: Detection of Multiclass Objects in Optical Remote Sensing Images
4.12.30. SB-MSN -> Sampling-Balance based Multi-stage Network (SB-MSN) for aerial image object detection. Code for 2021 paper: Improving Training Instance Quality in Aerial Image Object Detection With a Sampling-Balance-Based Multistage Network
4.12.31. yoltv5 -> detects objects in arbitrarily large aerial or satellite images that far exceed the ~600×600 pixel size typically ingested by deep learning object detection frameworks. Uses YOLOv5 & pytorch
4.12.32. AIR -> A deep learning object detector framework written in Python for supporting Land Search and Rescue Missions
4.12.33. dior_detect -> benchmarks for object detection on DIOR dataset
4.12.34. Panchromatic to Multispectral: Object Detection Performance as a Function of Imaging Bands -> Medium article, concludes that more bands are not always beneficial, but likely varies by use case
4.12.35. OPLD-Pytorch -> code for 2020 paper: Learning Point-Guided Localization for Detection in Remote Sensing Images
4.12.36. F3Net -> code for 2020 paper: Feature Fusion and Filtration Network for Object Detection in Optical Remote Sensing Images
4.12.37. GLNet -> code for 2021 paper: Global to Local: Clip-LSTM-Based Object Detection From Remote Sensing Images
4.12.38. SRAF-Net -> code for 2021 paper: A Scene-Relevant Anchor-Free Object Detection Network in Remote Sensing Images
4.12.39. object_detection_in_remote_sensing_images -> using CNN and attention mechanism
4.12.40. SHAPObjectDetection -> code for 2022 paper: SHAP-Based Interpretable Object Detection Method for Satellite Imagery
4.12.41. NWD -> code for 2021 paper: A Normalized Gaussian Wasserstein Distance for Tiny Object Detection. Uses AI-TOD dataset
4.12.42. MSFC-Net -> code for 2021 paper: Multiscale Semantic Fusion-Guided Fractal Convolutional Object Detection Network for Optical Remote Sensing Imagery
4.12.43. LO-Det -> code for the 2021 paper: LO-Det: Lightweight Oriented Object Detection in Remote Sensing Images
4.12.44. R2IPoints -> code for 2022 paper: R²IPoints: Pursuing Rotation-Insensitive Point Representation for Aerial Object Detection
4.12.45. Object-Detection -> code for 2022 paper: Multi-Scale Object Detection with the Pixel Attention Mechanism in a Complex Background
4.12.46. mmdet-rfla -> code for 2022 paper: RFLA: Gaussian Receptive based Label Assignment for Tiny Object Detection
4.12.47. Interactive-Multi-Class-Tiny-Object-Detection -> code for 2022 paper: Interactive Multi-Class Tiny-Object Detection
4.12.48. small-object-detection-benchmark -> code for ICIP 2022 paper: Slicing Aided Hyper Inference and Fine-tuning for Small Object Detection
4.12.49. OD-Satellite-iSAID -> Object Detection in Aerial Images: A Case Study on Performance Improvement using iSAID
4.12.50. Large-Selective-Kernel-Network -> code for 2023 paper: Large Selective Kernel Network for Remote Sensing Object Detection
4.12.51. Satellite_Imagery_Detection_YOLOV7 -> YOLOV7 applied to xView1 Dataset
4.12.52. FSANet -> code for 2022 paper: FSANet: Feature-and-Spatial-Aligned Network for Tiny Object Detection in Remote Sensing Images
5. Object counting
When the object count, but not its shape is required, U-net can be used to treat this as an image-to-image translation problem.
5.1. centroid-unet -> Centroid-UNet is deep neural network model to detect centroids from satellite images, with paper BEGINNER
5.2. cownter_strike -> counting cows, located with point-annotations, two models: CSRNet (a density-based method) & LCFCN (a detection-based method)
5.3. DO-U-Net -> an effective approach for when the size of an object needs to be known, as well as the number of objects in the image, initially created to segment and count Internally Displaced People (IDP) camps in Afghanistan
5.5. Counting from Sky -> A Large-scale Dataset for Remote Sensing Object Counting and A Benchmark Method
5.6. PSGCNet -> code for 2022 paper: PSGCNet: A Pyramidal Scale and Global Context Guided Network for Dense Object Counting in Remote Sensing Images
5.7. psgcnet -> code for 2022 paper: PSGCNet: A Pyramidal Scale and Global Context Guided Network for Dense Object Counting in Remote-Sensing Images
6. Regression
Regression prediction of windspeed.
Regression in remote sensing involves predicting continuous variables such as wind speed, tree height, or soil moisture from an image. Both classical machine learning and deep learning approaches can be used to accomplish this task. Classical machine learning utilizes feature engineering to extract numerical values from the input data, which are then used as input for a regression algorithm like linear regression. On the other hand, deep learning typically employs a convolutional neural network (CNN) to process the image data, followed by a fully connected neural network (FCNN) for regression. The FCNN is trained to map the input image to the desired output, providing predictions for the continuous variables of interest. Image source
6.1. python-windspeed -> Predicting windspeed of hurricanes from satellite images, uses CNN regression in keras
6.2. hurricane-wind-speed-cnn -> Predicting windspeed of hurricanes from satellite images, uses CNN regression in keras
6.3. GEDI-BDL -> code for paper: Global canopy height regression and uncertainty estimation from GEDI LIDAR waveforms with deep ensembles
6.4. Traffic density estimation as a regression problem instead of object detection -> inspired by paper: Traffic density estimation method from small satellite imagery: Towards frequent remote sensing of car traffic
6.5. OpticalWaveGauging_DNN -> code for 2020 paper: Optical wave gauging using deep neural networks
6.6. satellite-pose-estimation -> adapts a ResNet50 model architecture to perform pose estimation on several series of satellite images (both real and synthetic)
6.7. Tropical Cyclone Wind Estimation Competition -> on RadiantEarth MLHub
7. Cloud detection & removal
(left) False colour image and (right) a cloud & shadow mask.
Clouds are a major issue in remote sensing images as they can obscure the underlying ground features. This hinders the accuracy and effectiveness of remote sensing analysis, as the obscured regions cannot be properly interpreted. In order to address this challenge, various techniques have been developed to detect clouds in remote sensing images. Both classical algorithms and deep learning approaches can be employed for cloud detection. Classical algorithms typically use threshold-based techniques and hand-crafted features to identify cloud pixels. However, these techniques can be limited in their accuracy and are sensitive to changes in image appearance and cloud structure. On the other hand, deep learning approaches leverage the power of convolutional neural networks (CNNs) to accurately detect clouds in remote sensing images. These models are trained on large datasets of remote sensing images, allowing them to learn and generalize the unique features and patterns of clouds. The generated cloud mask can be used to identify the cloud pixels and eliminate them from further analysis or, alternatively, cloud inpainting techniques can be used to fill in the gaps left by the clouds. This approach helps to improve the accuracy of remote sensing analysis and provides a clearer view of the ground, even in the presence of clouds. Image adapted from this source
7.1. CloudSEN12 -> Sentinel 2 cloud dataset with a varierty of models here
7.2. From this article on sentinelhub there are three popular classical algorithms that detects thresholds in multiple bands in order to identify clouds. In the same article they propose using semantic segmentation combined with a CNN for a cloud classifier (excellent review paper here), but state that this requires too much compute resources.
7.3. This article compares a number of ML algorithms, random forests, stochastic gradient descent, support vector machines, Bayesian method.
7.4. Segmentation of Clouds in Satellite Images Using Deep Learning -> semantic segmentation using a Unet on the Kaggle 38-Cloud dataset
7.5. Cloud Detection in Satellite Imagery compares FPN+ResNet18 and CheapLab architectures on Sentinel-2 L1C and L2A imagery
7.6. Benchmarking Deep Learning models for Cloud Detection in Landsat-8 and Sentinel-2 images
7.7. Landsat-8 to Proba-V Transfer Learning and Domain Adaptation for Cloud detection
7.8. Multitemporal Cloud Masking in Google Earth Engine
7.9. s2cloudmask -> Sentinel-2 Cloud and Shadow Detection using Machine Learning
7.10. sentinel2-cloud-detector -> Sentinel Hub Cloud Detector for Sentinel-2 images in Python
7.11. dsen2-cr -> cloud removal in Sentinel-2 imagery using a deep residual neural network and SAR-optical data fusion, contains the model code, written in Python/Keras, as well as links to pre-trained checkpoints and the SEN12MS-CR dataset
7.12. pyatsa -> Python package implementing the Automated Time-Series Analysis method for masking clouds in satellite imagery developed by Zhu and Helmer 2018
7.13. decloud -> Decloud enables the training of various deep nets to remove clouds in optical image, using e.g. Sentinel 1 & 2
7.14. cloudless -> Deep learning pipeline for orbital satellite data for detecting clouds
7.15. Deep-Gapfill -> Official implementation of Optical image gap filling using deep convolutional autoencoder from optical and radar images
7.16. satellite-cloud-removal-dip -> Satellite cloud removal with Deep Image Prior, with paper
7.17. cloudFCN -> Python 3 package for Fully Convolutional Network development, specifically for cloud masking
7.18. Fmask -> Fmask (Function of mask) is used for automated clouds, cloud shadows, snow, and water masking for Landsats 4-9 and Sentinel 2 images, in Matlab. Also see PyFmask
7.20. cloud-cover-winners -> Code from the winning submissions for the On Cloud N: Cloud Cover Detection Challenge
7.21. On-Cloud-N: Cloud Cover Detection Challenge - 19th Place Solution
7.22. ukis-csmask -> package to masks clouds in Sentinel-2, Landsat-8, Landsat-7 and Landsat-5 images
7.23. OpenSICDR -> long list of satellite image cloud detection resources
7.24. RS-Net -> code for the paper: A cloud detection algorithm for satellite imagery based on deep learning
7.25. Clouds-Segmentation-Project -> treats as a 3 class problem; Open clouds, Closed clouds and no clouds, uses pytorch on a dataset that consists of IR & Visual Grayscale images
7.26. STGAN -> PyTorch Implementation of STGAN for Cloud Removal in Satellite Images, with paper
7.27. mcgan-cvprw2017-pytorch -> code for 2017 paper: Filmy Cloud Removal on Satellite Imagery with Multispectral Conditional Generative Adversarial Nets
7.28. Cloud-Net: A semantic segmentation CNN for cloud detection -> an end-to-end cloud detection algorithm for Landsat 8 imagery, trained on 38-Cloud Training Set
7.29. fcd -> code for 2021 paper: Fixed-Point GAN for Cloud Detection. A weakly-supervised approach, training with only image-level labels
7.30. CloudX-Net -> an efficient and robust architecture used for detection of clouds from satellite images
7.32. 38Cloud-Medium -> Walk-through using u-net to detect clouds in satellite images with fast.ai
7.33. cloud_detection_using_satellite_data -> performed on Sentinel 2 data
7.34. Luojia1-Cloud-Detection -> Luojia-1 Satellite Visible Band Nighttime Imagery Cloud Detection
7.35. SEN12MS-CR-TS -> code for 2022 paper: A Remote Sensing Data Set for Multi-modal Multi-temporal Cloud Removal
7.36. ES-CCGAN -> This is a dehazed method for remote sensing image, which based on CycleGAN
7.37. Cloud_Classification_DL -> Classifying cloud organization patterns from satellite images using Deep Learning techniques (Mask R-CNN)
7.38. CNN-based-Cloud-Detection-Methods -> Understanding the Role of Receptive Field of Convolutional Neural Network for Cloud Detection in Landsat 8 OLI Imagery
7.39. cloud-removal-deploy -> flask app for cloud removal
7.40. CloudMattingGAN -> code for 2019 paper: Generative Adversarial Training for Weakly Supervised Cloud Matting
7.41. atrain-cloudseg -> Official repository for the A-Train Cloud Segmentation Dataset
7.42. CDnet -> code for 2019 paper: CNN-Based Cloud Detection for Remote Sensing Imager
7.43. GLNET -> code for 2021 paper: Convolutional Neural Networks Based Remote Sensing Scene Classification under Clear and Cloudy Environments
7.44. CDnetV2 -> code for 2021 paper: CNN-Based Cloud Detection for Remote Sensing Imagery With Cloud-Snow Coexistence
7.45. grouped-features-alignment -> code for 2021 paper: Unsupervised Domain Adaptation for Cloud Detection Based on Grouped Features Alignment and Entropy Minimization
7.46. Detecting Cloud Cover Via Sentinel-2 Satellite Data -> blog post on Benjamin Warners Top-10 Percent Solution to DrivenData’s On CloudN Competition using fast.ai & customized version of XResNeXt50. Repo
7.47. AISD -> code (Matlab) and dataset for 2020 paper: Deeply supervised convolutional neural network for shadow detection based on a novel aerial shadow imagery dataset
7.48. CloudGAN -> Detecting and Removing Clouds from RGB-images using Image Inpainting
7.49. Using GANs to Augment Data for Cloud Image Segmentation Task -> code for 2021 paper
7.50. Cloud-Segmentation-from-Satellite-Imagery -> applied to Sentinel-2 dataset
7.51. HRC_WHU -> High-Resolution Cloud Detection Dataset comprising 150 RGB images and a resolution varying from 0.5 to 15 m in different global regions
7.52. MEcGANs -> Cloud Removal from Satellite Imagery using Multispectral Edge-filtered Conditional Generative Adversarial Networks
7.53. CloudXNet -> code for 2020 paper: CloudX-net: A robust encoder-decoder architecture for cloud detection from satellite remote sensing images
7.54. refined-unet-lite -> code for 2022 paper: Refined UNet Lite: End-to-End Lightweight Network for Edge-precise Cloud Detection
7.55. cloud-buster -> Sentinel-2 L1C and L2A Imagery with Fewer Clouds
7.56. SatelliteCloudGenerator -> A PyTorch-based tool to generate clouds for satellite images
7.57. SEnSeI -> A python 3 package for developing sensor independent deep learning models for cloud masking in satellite imagery
7.58. cloud-detection-venus -> code for 2022 paper: Using Convolutional Neural Networks for Cloud Detection on VENμS Images over Multiple Land-Cover Types
7.59. explaining_cloud_effects -> code for 2022 paper: Explaining the Effects of Clouds on Remote Sensing Scene Classification
7.60. Clouds-Images-Segmentation -> Marine Stratocumulus Cloud-Type Classification from SEVIRI Using Convolutional Neural Networks
7.61. DeCloud-GAN -> code for 2021 paper: DeCloud GAN: An Advanced Generative Adversarial Network for Removing Cloud Cover in Optical Remote Sensing Imagery
7.62. km_predict -> KappaMask, or km-predict, is a cloud detector for Sentinel-2 Level-1C and Level-2A input products applied to S2 full image prediction
8. Change detection
(left) Initial and (middle) after some development, with (right) the change highlighted.
Change detection is a vital component of remote sensing analysis, enabling the monitoring of landscape changes over time. This technique can be applied to identify a wide range of changes, including land use changes, urban development, coastal erosion, and deforestation. Change detection can be performed on a pair of images taken at different times, or by analyzing multiple images collected over a period of time. It is important to note that while change detection is primarily used to detect changes in the landscape, it can also be influenced by the presence of clouds and shadows. These dynamic elements can alter the appearance of the image, leading to false positives in change detection results. Therefore, it is essential to consider the impact of clouds and shadows on change detection analysis, and to employ appropriate methods to mitigate their influence. Image source
8.1. awesome-remote-sensing-change-detection lists many datasets and publications
8.2. Change-Detection-Review -> A review of change detection methods, including code and open data sets for deep learning
8.3. Change Detection using Siamese Networks -> Medium article BEGINNER
8.4. STANet -> official implementation of the spatial-temporal attention neural network (STANet) for remote sensing image change detection BEGINNER
8.5. UNet-based-Unsupervised-Change-Detection -> A convolutional neural network (CNN) and semantic segmentation is implemented to detect the changes between the images, as well as classify the changes into the correct semantic class, with arxiv paper BEGINNER
8.6. BIT_CD -> Official Pytorch Implementation of Remote Sensing Image Change Detection with Transformers
8.7. Unstructured-change-detection-using-CNN
8.8. Siamese neural network to detect changes in aerial images -> uses Keras and VGG16 architecture
8.9. Change Detection in 3D: Generating Digital Elevation Models from Dove Imagery
8.10. QGIS plugin for applying change detection algorithms on high resolution satellite imagery
8.11. LamboiseNet -> Master thesis about change detection in satellite imagery using Deep Learning
8.12. Fully Convolutional Siamese Networks for Change Detection -> with paper
8.13. Urban Change Detection for Multispectral Earth Observation Using Convolutional Neural Networks -> with paper, used the Onera Satellite Change Detection (OSCD) dataset
8.14. IAug_CDNet -> Official Pytorch Implementation of Adversarial Instance Augmentation for Building Change Detection in Remote Sensing Images
8.15. dpm-rnn-public -> Code implementing a damage mapping method combining satellite data with deep learning
8.16. SenseEarth2020-ChangeDetection -> 1st place solution to the Satellite Image Change Detection Challenge hosted by SenseTime; predictions of five HRNet-based segmentation models are ensembled, serving as pseudo labels of unchanged areas
8.17. KPCAMNet -> Python implementation of the paper Unsupervised Change Detection in Multi-temporal VHR Images Based on Deep Kernel PCA Convolutional Mapping Network
8.18. CDLab -> benchmarking deep learning-based change detection methods.
8.19. Siam-NestedUNet -> The pytorch implementation for "SNUNet-CD: A Densely Connected Siamese Network for Change Detection of VHR Images"
8.20. SUNet-change_detection -> Implementation of paper SUNet: Change Detection for Heterogeneous Remote Sensing Images from Satellite and UAV Using a Dual-Channel Fully Convolution Network
8.21. Self-supervised Change Detection in Multi-view Remote Sensing Images
8.22. MFPNet -> Remote Sensing Change Detection Based on Multidirectional Adaptive Feature Fusion and Perceptual Similarity
8.23. GitHub for the DIUx xView Detection Challenge -> The xView2 Challenge focuses on automating the process of assessing building damage after a natural disaster
8.24. DASNet -> Dual attentive fully convolutional siamese networks for change detection of high-resolution satellite images
8.25. Self-Attention for Raw Optical Satellite Time Series Classification
8.26. planet-movement -> Find and process Planet image pairs to highlight object movement
8.27. temporal-cluster-matching -> detecting change in structure footprints from time series of remotely sensed imagery
8.28. autoRIFT -> fast and intelligent algorithm for finding the pixel displacement between two images
8.29. DSAMNet -> Code for “A Deeply Supervised Attention Metric-Based Network and an Open Aerial Image Dataset for Remote Sensing Change Detection”. The main types of changes in the dataset include: (a) newly built urban buildings; (b) suburban dilation; (c) groundwork before construction; (d) change of vegetation; (e) road expansion; (f) sea construction.
8.30. SRCDNet -> The pytorch implementation for "Super-resolution-based Change Detection Network with Stacked Attention Module for Images with Different Resolutions ". SRCDNet is designed to learn and predict change maps from bi-temporal images with different resolutions
8.31. Land-Cover-Analysis -> Land Cover Change Detection using Satellite Image Segmentation
8.33. Satellite-Image-Alignment-Differencing-and-Segmentation -> thesis on change detection
8.34. Change Detection in Multi-temporal Satellite Images -> uses Principal Component Analysis (PCA) and K-means clustering
8.35. Unsupervised Change Detection Algorithm using PCA and K-Means Clustering -> in Matlab but has paper
8.36. ChangeFormer -> A Transformer-Based Siamese Network for Change Detection. Uses transformer architecture to address the limitations of CNN in handling multi-scale long-range details. Demonstrates that ChangeFormer captures much finer details compared to the other SOTA methods, achieving better performance on benchmark datasets
8.37. Heterogeneous_CD -> Heterogeneous Change Detection in Remote Sensing Images. Accompanies Code-Aligned Autoencoders for Unsupervised Change Detection in Multimodal Remote Sensing Images
8.38. ChangeDetectionProject -> Trying out Active Learning in with deep CNNs for Change detection on remote sensing data
8.39. DSFANet -> Unsupervised Deep Slow Feature Analysis for Change Detection in Multi-Temporal Remote Sensing Images
8.40. siamese-change-detection -> Targeted synthesis of multi-temporal remote sensing images for change detection using siamese neural networks
8.41. Bi-SRNet -> code for 2022 paper: Bi-Temporal Semantic Reasoning for the Semantic Change Detection in HR Remote Sensing Images
8.42. SiROC -> Implementation of the paper: Spatial Context Awareness for Unsupervised Change Detection in Optical Satellite Images. Applied to Sentinel-2 and high-resolution Planetscope imagery on four datasets
8.43. DSMSCN -> Tensorflow implementation for Change Detection in Multi-temporal VHR Images Based on Deep Siamese Multi-scale Convolutional Neural Networks
8.44. RaVAEn -> a lightweight, unsupervised approach for change detection in satellite data based on Variational Auto-Encoders (VAEs) with the specific purpose of on-board deployment. It flags changed areas to prioritise for downlink, shortening the response time
8.45. SemiCD -> Code for paper: Revisiting Consistency Regularization for Semi-supervised Change Detection in Remote Sensing Images. Achieves the performance of supervised CD even with access to as little as 10% of the annotated training data
8.46. FCCDN_pytorch -> code for paper: FCCDN: Feature Constraint Network for VHR Image Change Detection. Uses the LEVIR-CD building change detection dataset
8.47. INLPG_Python -> code for paper: Structure Consistency based Graph for Unsupervised Change Detection with Homogeneous and Heterogeneous Remote Sensing Images
8.48. NSPG_Python -> code for paper: Nonlocal patch similarity based heterogeneous remote sensing change detection
8.49. LGPNet-BCD -> code for 2021 paper: Building Change Detection for VHR Remote Sensing Images via Local-Global Pyramid Network and Cross-Task Transfer Learning Strategy
8.50. DS_UNet -> code for 2021 paper: Sentinel-1 and Sentinel-2 Data Fusion for Urban Change Detection using a Dual Stream U-Net, uses Onera Satellite Change Detection dataset
8.51. SiameseSSL -> code for 2022 paper: Urban change detection with a Dual-Task Siamese network and semi-supervised learning. Uses SpaceNet 7 dataset
8.52. CD-SOTA-methods -> Remote sensing change detection: State-of-the-art methods and available datasets
8.53. multimodalCD_ISPRS21 -> code for 2021 paper: Fusing Multi-modal Data for Supervised Change Detection
8.54. Unsupervised-CD-in-SITS-using-DL-and-Graphs -> code for article: Unsupervised Change Detection Analysis in Satellite Image Time Series using Deep Learning Combined with Graph-Based Approaches
8.55. LSNet -> code for 2022 paper: Extremely Light-Weight Siamese Network For Change Detection in Remote Sensing Image
8.56. Change-Detection-in-Remote-Sensing-Images -> using PCA & K-means
8.57. End-to-end-CD-for-VHR-satellite-image -> code for 2019 paper: End-to-End Change Detection for High Resolution Satellite Images Using Improved UNet++
8.58. Semantic-Change-Detection -> code for 2021 paper: SCDNET: A novel convolutional network for semantic change detection in high resolution optical remote sensing imagery
8.59. ERCNN-DRS_urban_change_monitoring -> code for 2021 paper: Neural Network-Based Urban Change Monitoring with Deep-Temporal Multispectral and SAR Remote Sensing Data
8.60. EGRCNN -> code for 2021 paper: Edge-guided Recurrent Convolutional Neural Network for Multi-temporal Remote Sensing Image Building Change Detection
8.61. Unsupervised-Remote-Sensing-Change-Detection -> code for 2021 paper: An Unsupervised Remote Sensing Change Detection Method Based on Multiscale Graph Convolutional Network and Metric Learning
8.62. CropLand-CD -> code for 2022 paper: A CNN-transformer Network with Multi-scale Context Aggregation for Fine-grained Cropland Change Detection
8.63. contrastive-surface-image-pretraining -> code for 2022 paper: Supervising Remote Sensing Change Detection Models with 3D Surface Semantics
8.64. dcvaVHROptical -> Deep Change Vector Analysis (DCVA) change detection. Code for 2019 paper: Unsupervised Deep Change Vector Analysis for Multiple-Change Detection in VHR Images
8.65. hyperdimensionalCD -> code for 2021 paper: Change Detection in Hyperdimensional Images Using Untrained Models
8.66. DSFANet -> code for 2018 paper: Unsupervised Deep Slow Feature Analysis for Change Detection in Multi-Temporal Remote Sensing Images
8.67. FCD-GAN-pytorch -> Fully Convolutional Change Detection Framework with Generative Adversarial Network (FCD-GAN) is a framework for change detection in multi-temporal remote sensing images
8.68. DARNet-CD -> code for 2022 paper: A Densely Attentive Refinement Network for Change Detection Based on Very-High-Resolution Bitemporal Remote Sensing Images
8.69. xView2_Vulcan -> Damage assessment using pre and post orthoimagery. Modified + productionized model based off the first-place model from the xView2 challenge.
8.70. ESCNet -> code for 2021 paper: An End-to-End Superpixel-Enhanced Change Detection Network for Very-High-Resolution Remote Sensing Images
8.71. ForestCoverChange -> Detecting and Predicting Forest Cover Change in Pakistani Areas Using Remote Sensing Imagery
8.72. deforestation-detection -> code for 2020 paper: DEEP LEARNING FOR HIGH-FREQUENCY CHANGE DETECTION IN UKRAINIAN FOREST ECOSYSTEM WITH SENTINEL-2
8.73. forest_change_detection -> forest change segmentation with time-dependent models, including Siamese, UNet-LSTM, UNet-diff, UNet3D models. Code for 2021 paper: Deep Learning for Regular Change Detection in Ukrainian Forest Ecosystem With Sentinel-2
8.74. SentinelClearcutDetection -> Scripts for deforestation detection on the Sentinel-2 Level-A images
8.75. clearcut_detection -> research & web-service for clearcut detection
8.76. CDRL -> code for 2022 paper: Unsupervised Change Detection Based on Image Reconstruction Loss
8.77. ddpm-cd -> code for 2022 paper: Remote Sensing Change Detection (Segmentation) using Denoising Diffusion Probabilistic Models
8.78. Remote-sensing-time-series-change-detection -> code for 2022 paper: Graph-based block-level urban change detection using Sentinel-2 time series
8.79. austin-ml-change-detection-demo -> A change detection demo for the Austin area using a pre-trained PyTorch model scaled with Dask on Planet imagery
8.80. dfc2021-msd-baseline -> A baseline for the "Multitemporal Semantic Change Detection" track of the 2021 IEEE GRSS Data Fusion Competition
8.81. CorrFusionNet -> code for 2020 paper: Multi-Temporal Scene Classification and Scene Change Detection with Correlation based Fusion
8.82. ChangeDetectionPCAKmeans -> MATLAB implementation for Unsupervised Change Detection in Satellite Images Using Principal Component Analysis and k-Means Clustering.
8.83. IRCNN -> code for 2022 paper: IRCNN: An Irregular-Time-Distanced Recurrent Convolutional Neural Network for Change Detection in Satellite Time Series
8.84. UTRNet -> An Unsupervised Time-Distance-Guided Convolutional Recurrent Network for Change Detection in Irregularly Collected Images
8.85. open-cd -> an open source change detection toolbox based on a series of open source general vision task tools
8.86. Tiny_model_4_CD -> code for 2022 paper: TINYCD: A (Not So) Deep Learning Model For Change Detection. Uses LEVIR-CD & WHU-CD datasets
8.87. FHD -> code for 2022 paper: Feature Hierarchical Differentiation for Remote Sensing Image Change Detection
8.88. Change detection with Raster Vision -> blog post with Colab notebook
8.89. building-expansion -> code for 2021 paper: Enhancing Environmental Enforcement with Near Real-Time Monitoring: Likelihood-Based Detection of Structural Expansion of Intensive Livestock Farms
8.90. SaDL_CD -> code for 2022 paper: Semantic-aware Dense Representation Learning for Remote Sensing Image Change Detection
8.91. EGCTNet_pytorch -> code for 2022 paper: Building Change Detection Based on an Edge-Guided Convolutional Neural Network Combined with a Transformer
8.92. S2-cGAN -> code for 2020 paper: S2-cGAN: Self-Supervised Adversarial Representation Learning for Binary Change Detection in Multispectral Images
8.93. A-loss-function-for-change-detection -> code for 2022 paper: UAL: Unchanged Area Loss-Function for Change Detection Networks
8.94. IEEE_TGRS_SSTFormer -> code for 2022 paper: Spectral–Spatial–Temporal Transformers for Hyperspectral Image Change Detection
8.95. DMINet -> code for 2023 paper: Change Detection on Remote Sensing Images Using Dual-Branch Multilevel Intertemporal Network
8.96. AFCF3D-Net -> code for 2023 paper: Adjacent-level Feature Cross-Fusion with 3D CNN for Remote Sensing Image Change Detection
8.97. DSAHRNet -> code for paper: A Deeply Attentive High-Resolution Network for Change Detection in Remote Sensing Images
8.98. RDPNet -> code for 2022 paper: RDP-Net: Region Detail Preserving Network for Change Detection
8.99. BGAAE_CD -> code for 2022 paper: Bipartite Graph Attention Autoencoders for Unsupervised Change Detection Using VHR Remote Sensing Images
8.100. Unsupervised-Change-Detection -> code for 2009 paper: Unsupervised Change Detection in Satellite Images Using Principal Component Analysis and k-Means Clustering
8.101. Metric-CD -> code for 2023 paper: Deep Metric Learning for Unsupervised Change Detection in Remote Sensing Images
8.102. HANet-CD -> code for 2023 paper: HANet: A hierarchical attention network for change detection with bi-temporal very-high-resolution remote sensing images
8.103. SRGCAE -> code for 2022 paper: Unsupervised Multimodal Change Detection Based on Structural Relationship Graph Representation Learning
8.104. change_detection_onera_baselines -> Siamese version of U-Net baseline model
8.105. SiamCRNN -> code for 2020 paper: Change Detection in Multisource VHR Images via Deep Siamese Convolutional Multiple-Layers Recurrent Neural Network
8.106. Graph-based methods for change detection in remote sensing images -> code for paper: Graph Learning Based on Signal Smoothness Representation for Homogeneous and Heterogeneous Change Detection
8.107. TransUNetplus2 -> code for 2023 paper: TransU-Net++: Rethinking attention gated TransU-Net for deforestation mapping
8.108. AR-CDNet -> code for 2023 paper: Towards Accurate and Reliable Change Detection of Remote Sensing Images via Knowledge Review and Online Uncertainty Estimation
8.109. CICNet -> code for 2023 paper: Compact Intertemporal Coupling Network for Remote Sensing Change Detection
8.110. BGINet -> code for 2023 paper: Remote Sensing Image Change Detection with Graph Interaction
9. Time series
Prediction of the next image in a series.
The analysis of time series observations in remote sensing data has numerous applications, including enhancing the accuracy of classification models and forecasting future patterns and events. Image source. Note: since classifying crops and predicting crop yield are such prominent use case for time series data, these tasks have dedicated sections after this one.
9.1. LANDSAT Time Series Analysis for Multi-temporal Land Cover Classification using Random Forest
9.2. temporalCNN -> Temporal Convolutional Neural Network for the Classification of Satellite Image Time Series
9.3. pytorch-psetae -> code for the paper: Satellite Image Time Series Classification with Pixel-Set Encoders and Temporal Self-Attention
9.4. satflow -> optical flow models for predicting future satellite images from current and past ones
9.5. esa-superresolution-forecasting -> Forecasting air pollution using ESA Sentinel-5p data, and an encoder-decoder convolutional LSTM neural network architecture, implemented in Pytorch
9.6. lightweight-temporal-attention-pytorch -> A PyTorch implementation of the Light Temporal Attention Encoder (L-TAE) for satellite image time series
9.7. dtwSat -> Time-Weighted Dynamic Time Warping for satellite image time series analysis
9.8. MTLCC -> code for paper: Multitemporal Land Cover Classification Network. A recurrent neural network approach to encode multi-temporal data for land cover classification
9.9. PWWB -> Code for the 2021 paper: Real-Time Spatiotemporal Air Pollution Prediction with Deep Convolutional LSTM through Satellite Image Analysis
9.10. spaceweather -> predicting geomagnetic storms from satellite measurements of the solar wind and solar corona, uses LSTMs
9.11. Forest_wildfire_spreading_convLSTM -> Modeling of the spreading of forest wildfire using a neural network with ConvLSTM cells. Prediction 3-days forward
9.12. ConvTimeLSTM -> Extension of ConvLSTM and Time-LSTM for irregularly spaced images, appropriate for Remote Sensing
9.13. dl-time-series -> Deep Learning algorithms applied to characterization of Remote Sensing time-series
9.14. tpe -> code for 2022 paper: Generalized Classification of Satellite Image Time Series With Thermal Positional Encoding
9.15. wildfire_forecasting -> code for 2021 paper: Deep Learning Methods for Daily Wildfire Danger Forecasting. Uses ConvLSTM
9.16. satellite_image_forecasting -> predict future satellite images from past ones using features such as precipitation and elevation maps. Entry for the EarthNet2021 challenge
9.17. Deep Learning for Cloud Gap-Filling on Normalized Difference Vegetation Index using Sentinel Time-Series -> A CNN-RNN based model that identifies correlations between optical and SAR data and exports dense Normalized Difference Vegetation Index (NDVI) time-series of a static 6-day time resolution and can be used for Events Detection tasks
9.18. DeepSatModels -> code for the 2023 paper: ViTs for SITS: Vision Transformers for Satellite Image Time Series
9.19. Presto -> code for the 2023 paper: Lightweight, Pre-trained Transformers for Remote Sensing Timeseries
10. Crop classification
(left) false colour image and (right) the crop map.
Crop classification in remote sensing is the identification and mapping of different crops in images or sequences of images. It aims to provide insight into the distribution and composition of crops in a specific area, with applications that include monitoring crop growth and evaluating crop damage. Both traditional machine learning methods, such as decision trees and support vector machines, and deep learning techniques, such as convolutional neural networks (CNNs), can be used to perform crop classification. The optimal method depends on the size and complexity of the dataset, the desired accuracy, and the available computational resources. However, the success of crop classification relies heavily on the quality and resolution of the input data, as well as the availability of labeled training data. Image source: High resolution satellite imaging sensors for precision agriculture by Chenghai Yang
10.1. Classification of Crop Fields through Satellite Image Time Series -> using a pytorch-psetae & Sentinel-2 data
10.2. CropDetectionDL -> using GRU-net, First place solution for Crop Detection from Satellite Imagery competition organized by CV4A workshop at ICLR 2020
10.3. Radiant-Earth-Spot-the-Crop-Challenge -> The main objective of this challenge was to use time-series of Sentinel-2 multi-spectral data to classify crops in the Western Cape of South Africa. The challenge was to build a machine learning model to predict crop type classes for the test dataset
10.4. Crop-Classification -> crop classification using multi temporal satellite images
10.5. DeepCropMapping -> A multi-temporal deep learning approach with improved spatial generalizability for dynamic corn and soybean mapping, uses LSTM
10.6. CropMappingInterpretation -> An interpretation pipeline towards understanding multi-temporal deep learning approaches for crop mapping
10.7. timematch -> code for 2022 paper: A method to perform unsupervised cross-region adaptation of crop classifiers trained with satellite image time series. We also introduce an open-access dataset for cross-region adaptation with SITS from four different regions in Europe
10.8. elects -> code for 2022 paper: End-to-End Learned Early Classification of Time Series for In-Season Crop Type Mapping
11. Crop yield
Wheat yield data. Blue vertical lines denote observation dates.
Crop yield is a crucial metric in agriculture, as it determines the productivity and profitability of a farm. It is defined as the amount of crops produced per unit area of land and is influenced by a range of factors including soil fertility, weather conditions, the type of crop grown, and pest and disease control. By utilizing time series of satellite images, it is possible to perform accurate crop type classification and take advantage of the seasonal variations specific to certain crops. This information can be used to optimize crop management practices and ultimately improve crop yield. However, to achieve accurate results, it is essential to consider the quality and resolution of the input data, as well as the availability of labeled training data. Appropriate pre-processing and feature extraction techniques must also be employed. Image source.
11.1. Crop yield Prediction with Deep Learning -> code for the paper Deep Gaussian Process for Crop Yield Prediction Based on Remote Sensing Data
11.2. Deep-Transfer-Learning-Crop-Yield-Prediction
11.3. Crop-Yield-Prediction-using-ML -> A simple Web application developed in order to provide the farmers/users an approximation on how much amount of crop yield will be produced depending upon the given input
11.4. Building a Crop Yield Prediction App in Senegal Using Satellite Imagery and Jupyter Voila
11.5. Crop Yield Prediction Using Deep Neural Networks and LSTM
11.6. Deep transfer learning techniques for crop yield prediction, published in COMPASS 2018
11.7. Understanding crop yield predictions from CNNs
11.8. Advanced Deep Learning Techniques for Predicting Maize Crop Yield using Sentinel-2 Satellite Imagery
11.9. pycrop-yield-prediction -> A PyTorch Implementation of Jiaxuan You's Deep Gaussian Process for Crop Yield Prediction
11.10. PredictYield -> using data scraped from Google Earth Engine, this predicts the yield of Corn, Soybean, and Wheat in the USA with Keras
11.11. Crop-Yield-Prediction-and-Estimation-using-Time-series-remote-sensing-data -> student research
11.12. Yield-Prediction-Using-Sentinel-Data
11.13. SPACY -> Satellite Prediction of Aggregate Corn Yield
11.14. cropyieldArticle -> code for 2022 paper: Scalable Crop Yield Prediction with Sentinel-2 Time Series and Temporal Convolutional Network
11.15. CNN-RNN-Yield-Prediction -> code for 2020 paper: A CNN-RNN Framework for Crop Yield Prediction
11.16. Yield-Prediction-DNN -> code for paper: Crop Yield Prediction Using Deep Neural Networks
12. Wealth and economic activity
COVID-19 impacts on human and economic activities.
The traditional approach of collecting economic data through ground surveys is a time-consuming and resource-intensive process. However, advancements in satellite technology and machine learning offer an alternative solution. By utilizing satellite imagery and applying machine learning algorithms, it is possible to obtain accurate and current information on economic activity with greater efficiency. This shift towards satellite imagery-based forecasting not only provides cost savings but also offers a wider and more comprehensive perspective of economic activity. As a result, it is poised to become a valuable asset for both policymakers and businesses. Image source.
12.1. Using publicly available satellite imagery and deep learning to understand economic well-being in Africa, Nature Comms 22 May 2020 -> Used CNN on Ladsat imagery (night & day) to predict asset wealth of African villages
12.2. Combining Satellite Imagery and machine learning to predict poverty -> review article
12.3. Measuring Human and Economic Activity from Satellite Imagery to Support City-Scale Decision-Making during COVID-19 Pandemic -> arxiv article
12.4. Predicting Food Security Outcomes Using CNNs for Satellite Tasking -> arxiv article
12.5. Measuring the Impacts of Poverty Alleviation Programs with Satellite Imagery and Deep Learning -> code and paper
12.6. Building a Spatial Model to Classify Global Urbanity Levels -> estimage global urbanity levels from population data, nightime lights and road networks
12.7. deeppop -> Deep Learning Approach for Population Estimation from Satellite Imagery, also on Github
12.8. Estimating telecoms demand in areas of poor data availability -> with papers on arxiv and Science Direct
12.9. satimage -> Code and models for the manuscript "Predicting Poverty and Developmental Statistics from Satellite Images using Multi-task Deep Learning". Predict the main material of a roof, source of lighting and source of drinking water for properties, from satellite imagery
12.10. africa_poverty -> Using publicly available satellite imagery and deep learning to understand economic well-being in Africa
12.11. Predicting-Poverty -> Combining satellite imagery and machine learning to predict poverty, in PyTorch
12.12. income-prediction -> Predicting average yearly income based on satellite imagery using CNNs, uses pytorch
12.13. urban_score -> Pytorch Implementation of paper: Learning to score economic development from satellite imagery
12.14. READ -> Pytorch Implementation of paper: Lightweight and robust representation of economic scales from satellite imagery
12.15. Slum-classification -> Binary classification on a very high-resolution satellite image in case of mapping informal settlements using unet
12.16. Predicting_Poverty -> uses daytime & luminosity of nighttime satellite images
12.17. Cancer-Prevalence-Satellite-Images -> Predict Health Outcomes from Features of Satellite Images
12.18. Mapping Poverty in Bangladesh with Satellite Images and Deep Learning -> combines health data with OpenStreetMaps Data & night and daytime satellite imagery
12.19. Population Estimation from Satellite Imagery
12.20. Deep_Learning_Satellite_Imd -> code for "Project Bhoomi" - Using Deep Learning on Satellite Imagery to predict population and economic indicators
12.21. satellite_led_liverpool -> code for 2017 paper: Remote Sensing-Based Measurement of Living Environment Deprivation - Improving Classical Approaches with Machine Learning
12.22. Predicting_Energy_Consumption_With_Convolutional_Neural_Networks
12.23. SustainBench -> Benchmarks for Monitoring the Sustainable Development Goals with Machine Learning
13. Disaster response
Detecting buildings destroyed in a disaster.
Remote sensing images are used in disaster response to identify and assess damage to an area. This imagery can be used to detect buildings that are damaged or destroyed, identify roads and road networks that are blocked, determine the size and shape of a disaster area, and identify areas that are at risk of flooding. Remote sensing images can also be used to detect and monitor the spread of forest fires and monitor vegetation health. Also checkout the sections on change detection and water/fire/building segmentation. Image source.
13.1. DisaVu -> combines building & damage detection and provides an app for viewing predictions
13.2. Soteria -> uses machine learning with satellite imagery to map natural disaster impacts for faster emergency response
13.3. DisasterHack -> Wildfire Mitigation: Computer Vision Identification of Hazard Fuels Using Landsat
13.4. forestcasting -> Forest fire prediction powered by analytics
13.5. Machine Learning-based Damage Assessment for Disaster Relief on Google AI blog -> uses object detection to locate buildings, then a classifier to determine if a building is damaged. Challenge of generalising due to small dataset
13.6. hurricane_damage -> Post-hurricane structure damage assessment based on aerial imagery with CNN
13.7. rescue -> code of the paper: Attention to fires: multi-channel deep-learning models forwildfire severity prediction
13.8. Disaster-Classification -> A disaster classification model to predict the type of disaster given an input image
13.9. Coarse-to-fine weakly supervised learning method for green plastic cover segmentation -> with paper
13.10. Detection of destruction in satellite imagery
13.11. BDD-Net -> code for 2020 paper: A General Protocol for Mapping Buildings Damaged by a Wide Range of Disasters Based on Satellite Imagery
13.12. Automatic_Disaster_Detection -> detect the affected area by natural disasters by using the way of semantic segmentation and change detection method
13.14. IBM-Disaster-Response-Hack -> identifying optimal terrestrial routes through calamity-stricken areas. Satellite image data informs road condition assessment and obstruction detection
13.15. Automatic Damage Annotation on Post-Hurricane Satellite Imagery -> detect damaged buildings using tensorflow object detection API. With repos here and here
13.16. Hurricane-Damage-Detection -> Waterloo's Hack the North 2020++ submission. A convolutional neural network model used to detect hurricane damage in RGB satellite images
13.17. wildfire_forecasting -> code for 2021 paper: Deep Learning Methods for Daily Wildfire Danger Forecasting. Uses ConvLSTM
13.18. Satellite Image Analysis with fast.ai for Disaster Recovery
13.19. shackleton -> leverages remote sensing imagery and machine learning techniques to provide insights into various transportation and evacuation scenarios in an interactive dashboard that conducts real-time computation
13.20. ai-vegetation-fuel -> Predicting Fuel Load from earth observation data using Machine Learning, using LightGBM & CatBoost
13.21. AI Helps Detect Disaster Damage From Satellite Imagery -> NVIDIA blog post
13.22. Turkey-Earthquake-2023-Building-Change-Detection -> The repository contains building footprints derived from Maxar open data imagery and change detection results by blackshark-ai
13.23. MS4D-Net-Building-Damage-Assessment -> code for 2022 paper: MS4D-Net: Multitask-Based Semi-Supervised Semantic Segmentation Framework with Perturbed Dual Mean Teachers for Building Damage Assessment from High-Resolution Remote Sensing Imagery
13.24. DAHiTra -> code for 2022 paper: Large-scale Building Damage Assessment using a Novel Hierarchical Transformer Architecture on Satellite Images. Uses xView2 xBD dataset
13.25. building-segmentation-disaster-resilience -> 2nd place solution in the Open Cities AI Challenge: Segmenting Buildings for Disaster Resilience
14. Super-resolution
Super resolution using multiple low resolution images as input.
Super-resolution is a technique aimed at improving the resolution of an imaging system. This process can be applied prior to other image processing steps to increase the visibility of small objects or boundaries. Despite its potential benefits, the use of super-resolution is controversial due to the possibility of introducing artifacts that could be mistaken for real features. Super-resolution techniques are broadly categorized into two groups: single image super-resolution (SISR) and multi-image super-resolution (MISR). SISR focuses on enhancing the resolution of a single image, while MISR utilizes multiple images of the same scene to create a high-resolution output. Each approach has its own advantages and limitations, and the choice of method depends on the specific application and desired outcome. Image source.
14.1. Multi image super-resolution (MISR)
Note that nearly all the MISR publications resulted from the PROBA-V Super Resolution competition
14.1.1. deepsum -> Deep neural network for Super-resolution of Unregistered Multitemporal images (ESA PROBA-V challenge)
14.1.2. 3DWDSRNet -> code to reproduce Satellite Image Multi-Frame Super Resolution (MISR) Using 3D Wide-Activation Neural Networks
14.1.3. RAMS -> Official TensorFlow code for paper Multi-Image Super Resolution of Remotely Sensed Images Using Residual Attention Deep Neural Networks
14.1.4. TR-MISR -> Transformer-based MISR framework for the the PROBA-V super-resolution challenge. With paper
14.1.5. HighRes-net -> Pytorch implementation of HighRes-net, a neural network for multi-frame super-resolution, trained and tested on the European Space Agency’s Kelvin competition
14.1.6. ProbaVref -> Repurposing the Proba-V challenge for reference-aware super resolution
14.1.7. The missing ingredient in deep multi-temporal satellite image super-resolution -> Permutation invariance harnesses the power of ensembles in a single model, with repo piunet
14.1.8. MSTT-STVSR -> Space-time Super-resolution for Satellite Video: A Joint Framework Based on Multi-Scale Spatial-Temporal Transformer, JAG, 2022
14.1.9. Self-Supervised Super-Resolution for Multi-Exposure Push-Frame Satellites
14.1.10. DDRN -> Deep Distillation Recursive Network for Video Satellite Imagery Super-Resolution
14.1.11. worldstrat -> SISR and MISR implementations of SRCNN
14.1.12. MISR-GRU -> Pytorch implementation of MISR-GRU, a deep neural network for multi image super-resolution (MISR), for ProbaV Super Resolution Competition
14.1.13. MSDTGP -> code for 2021 paper: Satellite Video Super-Resolution via Multiscale Deformable Convolution Alignment and Temporal Grouping Projection
14.1.14. proba-v-super-resolution-challenge -> Solution to ESA's satellite imagery super resolution challenge
14.1.15. PROBA-V-Super-Resolution -> solution using a custom deep learning architecture
14.2. Single image super-resolution (SISR)
14.2.1. Super Resolution for Satellite Imagery - srcnn repo
14.2.3. Random Forest Super-Resolution (RFSR repo) including sample data
14.2.4. Enhancing Sentinel 2 images by combining Deep Image Prior and Decrappify. Repo for deep-image-prior and article on decrappify
14.2.5. Image Super-Resolution using an Efficient Sub-Pixel CNN -> the keras docs have a great tutorial on this light weight but well performing model
14.2.6. super-resolution-using-gan -> Super-Resolution of Sentinel-2 Using Generative Adversarial Networks
14.2.7. Super-resolution of Multispectral Satellite Images Using Convolutional Neural Networks with paper
14.2.8. Multi-temporal Super-Resolution on Sentinel-2 Imagery using HighRes-Net, repo
14.2.9. SSPSR-Pytorch -> A spatial-spectral prior deep network for single hyperspectral image super-resolution
14.2.10. Sentinel-2 Super-Resolution: High Resolution For All (Bands)
14.2.11. CinCGAN -> Unofficial Implementation of Unsupervised Image Super-Resolution using Cycle-in-Cycle Generative Adversarial Networks
14.2.12. Satellite-image-SRGAN using PyTorch
14.2.13. EEGAN -> Edge Enhanced GAN For Remote Sensing Image Super-Resolution, TensorFlow 1.1
14.2.14. PECNN -> A Progressively Enhanced Network for Video Satellite Imagery Super-Resolution, minimal documentation
14.2.15. hs-sr-tvtv -> Enhanced Hyperspectral Image Super-Resolution via RGB Fusion and TV-TV Minimization
14.2.16. sr4rs -> Super resolution for remote sensing, with pre-trained model for Sentinel-2, SRGAN-inspired
14.2.17. Restoring old aerial images with Deep Learning -> Medium article on Super Resolution with Perceptual Loss function and real images as input
14.2.18. RFSR_TGRS -> code for the paper Hyperspectral Image Super-Resolution via Recurrent Feedback Embedding and Spatial-Spectral Consistency Regularization
14.2.19. SEN2VENµS -> a dataset for the training of Sentinel-2 super-resolution algorithms. With paper
14.2.20. TransENet -> code for 2021 paper: Transformer-based Multi-Stage Enhancement for Remote Sensing Image Super-Resolution
14.2.21. SG-FBGAN -> code for 2020 paper: Remote Sensing Image Super-Resolution via Saliency-Guided Feedback GANs
14.2.22. finetune_ESRGAN -> finetune the ESRGAN super resolution generator for remote sensing images and video
14.2.23. MIP -> code for 2021 paper: Unsupervised Remote Sensing Super-Resolution via Migration Image Prior
14.2.24. Optical-RemoteSensing-Image-Resolution -> code for 2018 paper: Deep Memory Connected Neural Network for Optical Remote Sensing Image Restoration. Two applications: Gaussian image denoising and single image super-resolution
14.2.25. HSENet -> code for 2021 paper: Hybrid-Scale Self-Similarity Exploitation for Remote Sensing Image Super-Resolution
14.2.26. SR_RemoteSensing -> Super-Resolution deep learning models for remote sensing data based on BasicSR
14.2.27. RSI-Net -> code for 2022 paper: A Deep Multi-task Convolutional Neural Network for Remote Sensing Image Super-resolution and Colorization
14.2.28. EDSR-Super-Resolution -> EDSR model using PyTorch applied to satellite imagery
14.2.29. CycleCNN -> code for 2021 paper: Nonpairwise-Trained Cycle Convolutional Neural Network for Single Remote Sensing Image Super-Resolution
14.2.30. SISR with with Real-World Degradation Modeling -> code for 2022 paper: Single-Image Super Resolution of Remote Sensing Images with Real-World Degradation Modeling
14.2.31. pixel-smasher -> code for 2020 paper: Super-Resolution Surface Water Mapping on the Canadian Shield Using Planet CubeSat Images and a Generative Adversarial Network
14.2.32. satellite-image-super-resolution -> A Comparative Study on CNN-Based Single-Image Super-Resolution Techniques for Satellite Images
14.2.33. SatelliteSR -> comparison of a number of techniques on the DOTA dataset
14.2.34. Image-Super-Resolution -> Super resolution RESNET network
14.2.35. Unsupervised Super Resolution for Sentinel-2 satellite imagery -> using Deep Image Prior (DIP), Zero-Shot Super Resolution (ΖSSR) & Degradation-Aware Super Resolution (DASR)
14.2.36. Spectral Super-Resolution of Satellite Imagery with Generative Adversarial Networks
14.2.37. Super resolution using GAN / 4x Improvement -> applied to Sentinel 2
14.2.38. rs-esrgan -> code for paper: RS-ESRGAN: Super-Resolution of Sentinel-2 Imagery Using Generative Adversarial Networks
14.2.39. TS-RSGAN -> code for paper: Super-Resolution of Remote Sensing Images for ×4 Resolution without Reference Images. Applied to Sentinel-2
14.2.40. CDCR -> code for 2023 paper: Combining Discrete and Continuous Representation: Scale-Arbitrary Super-Resolution for Satellite Images
14.3. Super-resolution - Miscellaneous
14.3.1. The value of super resolution — real world use case -> Medium article on parcel boundary detection with super-resolved satellite imagery
14.3.2. Super-Resolution on Satellite Imagery using Deep Learning -> Nov 2016 blog post by CosmiQ Works with a nice introduction to the topic. Proposes and demonstrates a new architecture with perturbation layers with practical guidance on the methodology and code. Three part series
14.3.3. Introduction to spatial resolution
14.3.4. Awesome-Super-Resolution -> another 'awesome' repo, getting a little out of date now
14.3.5. Super-Resolution (python) Utilities for managing large satellite images
14.3.6. pytorch-enhance -> Library of Image Super-Resolution Models, Datasets, and Metrics for Benchmarking or Pretrained Use. Also checkout this implementation in Jax
14.3.7. Super Resolution in OpenCV
14.3.8. AI-based Super resolution and change detection to enforce Sentinel-2 systematic usage -> Worldview-2 images (2m) were used to create a reference dataset and increase the spatial resolution of the Copernicus sensor from 10m to 5m
14.3.9. SRCDNet -> The pytorch implementation for "Super-resolution-based Change Detection Network with Stacked Attention Module for Images with Different Resolutions ". SRCDNet is designed to learn and predict change maps from bi-temporal images with different resolutions
14.3.10. Model-Guided Deep Hyperspectral Image Super-resolution -> code accompanying the paper: Model-Guided Deep Hyperspectral Image Super-Resolution
14.3.11. Super-resolving beyond satellite hardware -> paper assessing SR performance in reconstructing realistically degraded satellite images
14.3.12. satellite-pixel-synthesis-pytorch -> PyTorch implementation of NeurIPS 2021 paper: Spatial-Temporal Super-Resolution of Satellite Imagery via Conditional Pixel Synthesis
14.3.13. SRE-HAN -> Squeeze-and-Residual-Excitation Holistic Attention Network improves super-resolution (SR) on remote-sensing imagery compared to other state-of-the-art attention-based SR models
14.3.14. satsr -> A project to perform super-resolution on multispectral images from any satellite, including Sentinel 2, Landsat 8, VIIRS &MODIS
14.3.15. OLI2MSI -> dataset for remote sensing imagery super-resolution composed of Landsat8-OLI and Sentinel2-MSI images
14.3.16. MMSR -> Learning Mutual Modulation for Self-Supervised Cross-Modal Super-Resolution
14.3.17. HSRnet -> code for the 2021 paper: Hyperspectral Image Super-resolution via Deep Spatio-spectral Attention Convolutional Neural Networks
14.3.18. RRSGAN -> code for 2021 paper: RRSGAN: Reference-Based Super-Resolution for Remote Sensing Image
14.3.19. HDR-DSP-SR -> code for 2021 paper: Self-supervised multi-image super-resolution for push-frame satellite images
14.3.20. GAN-HSI-SR -> code for 2020 paper: Hyperspectral Image Super-Resolution by Band Attention Through Adversarial Learning
14.3.21. Restoring old aerial images with Deep Learning -> Medium article Super Resolution with Perceptual Loss function and real images as input
15. Pansharpening
Pansharpening example with a resolution difference of factor 4.
Pansharpening is a data fusion method that merges the high spatial detail from a high-resolution panchromatic image with the rich spectral information from a lower-resolution multispectral image. The result is a single, high-resolution color image that retains both the sharpness of the panchromatic band and the color information of the multispectral bands. This process enhances the spatial resolution while preserving the spectral qualities of the original images. Image source
15.1. Several algorithms described in the ArcGIS docs, with the simplest being taking the mean of the pan and RGB pixel value.
15.2. For into to classical methods see this notebook and this kaggle kernel
15.3. rio-pansharpen -> pansharpening Landsat scenes
15.4. Simple-Pansharpening-Algorithms
15.5. Working-For-Pansharpening -> long list of pansharpening methods and update of Awesome-Pansharpening
15.6. PSGAN -> A Generative Adversarial Network for Remote Sensing Image Pan-sharpening, arxiv paper
15.7. Pansharpening-by-Convolutional-Neural-Network
15.8. PBR_filter -> {P}ansharpening by {B}ackground {R}emoval algorithm for sharpening RGB images
15.9. py_pansharpening -> multiple algorithms implemented in python
15.10. Deep-Learning-PanSharpening -> deep-learning based pan-sharpening code package, we reimplemented include PNN, MSDCNN, PanNet, TFNet, SRPPNN, and our purposed network DIPNet
15.11. HyperTransformer -> A Textural and Spectral Feature Fusion Transformer for Pansharpening
15.12. DIP-HyperKite -> Hyperspectral Pansharpening Based on Improved Deep Image Prior and Residual Reconstruction
15.13. D2TNet -> code for 2022 paper: A ConvLSTM Network with Dual-direction Transfer for Pan-sharpening
15.14. PanColorGAN-VHR-Satellite-Images -> code for 2020 paper: Rethinking CNN-Based Pansharpening: Guided Colorization of Panchromatic Images via GANs
15.15. MTL_PAN_SEG -> code for 2019 paper: Multi-task deep learning for satellite image pansharpening and segmentation
15.16. Z-PNN -> code for 2022 paper: Pansharpening by convolutional neural networks in the full resolution framework
15.17. GTP-PNet -> code for 2021 paper: GTP-PNet: A residual learning network based on gradient transformation prior for pansharpening
15.18. UDL -> code for 2021 paper: Dynamic Cross Feature Fusion for Remote Sensing Pansharpening
15.19. PSData -> A Large-Scale General Pan-sharpening DataSet, which contains PSData3 (QB, GF-2, WV-3) and PSData4 (QB, GF-1, GF-2, WV-2).
15.20. AFPN -> Adaptive Detail Injection-Based Feature Pyramid Network For Pan-sharpening
15.21. pan-sharpening -> multiple methods demonstrated for multispectral and panchromatic images
15.22. PSGan-Family -> code for 2020 paper: PSGAN: A Generative Adversarial Network for Remote Sensing Image Pan-Sharpening
15.23. PanNet-Landsat -> code for 2017 paper: A Deep Network Architecture for Pan-Sharpening
15.24. DLPan-Toolbox -> code for 2022 paper: Machine Learning in Pansharpening: A Benchmark, from Shallow to Deep Networks
15.25. LPPN -> code for 2021 paper: Laplacian pyramid networks: A new approach for multispectral pansharpening
15.26. S2_SSC_CNN -> code for 2020 paper: Zero-shot Sentinel-2 Sharpening Using A Symmetric Skipped Connection Convolutional Neural Network
15.27. S2S_UCNN -> code for 2021 paper: Sentinel 2 sharpening using a single unsupervised convolutional neural network with MTF-Based degradation model
15.28. SSE-Net -> code for 2022 paper: Spatial and Spectral Extraction Network With Adaptive Feature Fusion for Pansharpening
15.29. UCGAN -> code for 2022 paper: Unsupervised Cycle-consistent Generative Adversarial Networks for Pan-sharpening
15.30. GCPNet -> code for 2022 paper: When Pansharpening Meets Graph Convolution Network and Knowledge Distillation
15.31. PanFormer -> code for 2022 paper: PanFormer: a Transformer Based Model for Pan-sharpening
15.32. Pansharpening -> code for 2021 paper: Pansformers: Transformer-Based Self-Attention Network for Pansharpening
15.33. Sentinel-2 Band Pan-Sharpening
15.34. PGCU -> code for 2023 paper: Probability-based Global Cross-modal Upsampling for Pansharpening
16. Image-to-image translation
(left) Sentinel-1 SAR input, (middle) translated to RGB and (right) Sentinel-2 true RGB image for comparison.
Image-to-image translation is a crucial aspect of computer vision that utilizes machine learning models to transform an input image into a new, distinct output image. In the field of remote sensing, it plays a significant role in bridging the gap between different imaging domains, such as converting Synthetic Aperture Radar (SAR) images into RGB (Red Green Blue) images. This technology has a wide range of applications, including improving image quality, filling in missing information, and facilitating cross-domain image analysis and comparison. By leveraging deep learning algorithms, image-to-image translation has become a powerful tool in the arsenal of remote sensing researchers and practitioners. Image source
16.1. How to Develop a Pix2Pix GAN for Image-to-Image Translation -> how to develop a Pix2Pix model for translating satellite photographs to Google map images. A good intro to GANS
16.2. A growing problem of ‘deepfake geography’: How AI falsifies satellite images
16.3. Kaggle Pix2Pix Maps -> dataset for pix2pix to take a google map satellite photo and build a street map
16.4. guided-deep-decoder -> With guided deep decoder, you can solve different image pair fusion problems, allowing super-resolution, pansharpening or denoising
16.5. hackathon-ci-2020 -> generate nighttime imagery from infrared observations
16.6. satellite-to-satellite-translation -> VAE-GAN architecture for unsupervised image-to-image translation with shared spectral reconstruction loss. Model is trained on GOES-16/17 and Himawari-8 L1B data
16.7. Pytorch implementation of UNet for converting aerial satellite images into google maps kinda images
16.8. Seamless-Satellite-image-Synthesis -> generate abitrarily large RGB images from a map
16.9. How to Develop a Pix2Pix GAN for Image-to-Image Translation -> article on machinelearningmastery.com
16.10. Satellite-Imagery-to-Map-Translation-using-Pix2Pix-GAN-framework
16.11. RSIT_SRM_ISD -> PyTorch implementation of Remote sensing image translation via style-based recalibration module and improved style discriminator
16.12. pix2pix_google_maps -> Converts satellite images to map images using pix2pix models
16.13. sar2color-igarss2018-chainer -> code for 2018 paper: Image Translation Between Sar and Optical Imagery with Generative Adversarial Nets
16.14. HSI2RGB -> Create realistic looking RGB images using remote sensing hyperspectral images
16.15. sat_to_map -> Learning mappings to generate city maps images from corresponding satellite images
16.16. pix2pix-GANs -> Generate Map using Satellite Image & PyTorch
17. Data fusion
Illustration of a fusion workflow.
Data fusion is a technique for combining information from different sources such as Synthetic Aperture Radar (SAR), optical imagery, and non-imagery data such as Internet of Things (IoT) sensor data. The integration of diverse data sources enables data fusion to overcome the limitations of individual sources, leading to the creation of models that are more accurate and informative than those constructed from a single source. Image source
17.1. Awesome-Data-Fusion-for-Remote-Sensing
17.2. UDALN_GRSL -> Deep Unsupervised Blind Hyperspectral and Multispectral Data Fusion
17.3. CropTypeMapping -> Crop type mapping from optical and radar (Sentinel-1&2) time series using attention-based deep learning
17.4. Multimodal-Remote-Sensing-Toolkit -> uses Hyperspectral and LiDAR Data
17.5. Aerial-Template-Matching -> development of an algorithm for template Matching on aerial imagery applied to UAV dataset
17.6. DS_UNet -> code for 2021 paper: Sentinel-1 and Sentinel-2 Data Fusion for Urban Change Detection using a Dual Stream U-Net, uses Onera Satellite Change Detection dataset
17.7. DDA_UrbanExtraction -> Unsupervised Domain Adaptation for Global Urban Extraction using Sentinel-1 and Sentinel-2 Data
17.8. swinstfm -> code for paper: Remote Sensing Spatiotemporal Fusion using Swin Transformer
17.9. LoveCS -> code for 2022 paper: Cross-sensor domain adaptation for high-spatial resolution urban land-cover mapping: from airborne to spaceborne imagery
17.10. comingdowntoearth -> code for 2021 paper: Implementation of 'Coming Down to Earth: Satellite-to-Street View Synthesis for Geo-Localization'
17.11. Matching between acoustic and satellite images
17.12. MapRepair -> Deep Cadastre Maps Alignment and Temporal Inconsistencies Fix in Satellite Images
17.13. Compressive-Sensing-and-Deep-Learning-Framework -> Compressive Sensing is used as an initial guess to combine data from multiple sources, with LSTM used to refine the result
17.14. DeepSim -> code for paper: DeepSIM: GPS Spoofing Detection on UAVs using Satellite Imagery Matching
17.15. MHF-net -> code for 2019 paper: Multispectral and Hyperspectral Image Fusion by MS/HS Fusion Net
17.16. Remote_Sensing_Image_Fusion -> code for 2021 paper: Semi-Supervised Remote Sensing Image Fusion Using Multi-Scale Conditional Generative Adversarial network with Siamese Structure
17.17. CNNs for Multi-Source Remote Sensing Data Fusion -> code for 2021 paper: Single-stream CNN with Learnable Architecture for Multi-source Remote Sensing Data
17.18. Deep Generative Reflectance Fusion -> Achieving Landsat-like reflectance at any date by fusing Landsat and MODIS surface reflectance with deep generative models
17.19. IEEE_TGRS_MDL-RS -> code for 2021 paper: More Diverse Means Better: Multimodal Deep Learning Meets Remote-Sensing Imagery Classification
17.20. SSRNET -> code for 2022 paper: SSR-NET: Spatial-Spectral Reconstruction Network for Hyperspectral and Multispectral Image Fusion
17.21. cross-view-image-matching -> code for 2019 paper: Bridging the Domain Gap for Ground-to-Aerial Image Matching
17.22. CoF-MSMG-PCNN -> code for 2020 paper: Remote Sensing Image Fusion via Boundary Measured Dual-Channel PCNN in Multi-Scale Morphological Gradient Domain
17.23. robust_matching_network_on_remote_sensing_imagery_pytorch -> code for 2019 paper: A Robust Matching Network for Gradually Estimating Geometric Transformation on Remote Sensing Imagery
17.24. edcstfn -> code for 2019 paper: An Enhanced Deep Convolutional Model for Spatiotemporal Image Fusion
17.25. ganstfm -> code for 2021 paper: A Flexible Reference-Insensitive Spatiotemporal Fusion Model for Remote Sensing Images Using Conditional Generative Adversarial Network
17.26. CMAFF -> code for 2021 paper: Cross-Modality Attentive Feature Fusion for Object Detection in Multispectral Remote Sensing Imagery
17.27. SOLC -> code for 2022 paper: MCANet: A joint semantic segmentation framework of optical and SAR images for land use classification. Uses WHU-OPT-SAR-dataset
17.28. MFT -> code for 2022 paper: Multimodal Fusion Transformer for Remote Sensing Image Classification
17.29. ISPRS_S2FL -> code for 2021 paper: Multimodal Remote Sensing Benchmark Datasets for Land Cover Classification with A Shared and Specific Feature Learning Model
17.30. HSHT-Satellite-Imagery-Synthesis -> code for thesis - Improving Flood Maps by Increasing the Temporal Resolution of Satellites Using Hybrid Sensor Fusion
17.31. MDC -> code for 2021 paper: Unsupervised Data Fusion With Deeper Perspective: A Novel Multisensor Deep Clustering Algorithm
17.32. FusAtNet -> code for 2020 paper: FusAtNet: Dual Attention based SpectroSpatial Multimodal Fusion Network for Hyperspectral and LiDAR Classification
17.33. AMM-FuseNet -> code for 2022 paper: AMM-FuseNet: Attention-Based Multi-Modal Image Fusion Network for Land Cover Mapping
17.34. MANet -> code for 2022 paper: MANet: A Network Architecture for Remote Sensing Spatiotemporal Fusion Based on Multiscale and Attention Mechanisms
17.35. DCSA-Net -> code for 2022 paper: Dynamic Convolution Self-Attention Network for Land-Cover Classification in VHR Remote-Sensing Images
17.36. deforestation-from-data-fusion -> Fusing Sentinel-1 and Sentinel-2 images for deforestation detection in the Brazilian Amazon under diverse cloud conditions
17.37. sct-fusion -> code for 2023 paper: Transformer-based Multi-Modal Learning for Multi Label Remote Sensing Image Classification
18. Generative Adversarial Networks (GANs)
Example generated images using a GAN.
Generative Adversarial Networks (GANs) are a type of deep learning architecture that leverages the power of competition between two neural networks. The objective of a GAN is to generate new, synthetic data that appears similar to real-world data. This is achieved by training the two networks, the generator and the discriminator, in a zero-sum game, where the generator attempts to produce data that is indistinguishable from the real data, while the discriminator tries to distinguish between the generated data and the real data. In the field of remote sensing, GANs have found numerous applications, particularly in generating synthetic data. This synthetic data can be used for a wide range of purposes, including data augmentation, data imbalance correction, and filling in missing or corrupted data. By generating realistic synthetic data, GANs can improve the performance of remote sensing algorithms and models, leading to more accurate and reliable results. Additionally, GANs can also be used for various other tasks in remote sensing, such as super-resolution, denoising, and inpainting. Image source
18.1. Using Generative Adversarial Networks to Address Scarcity of Geospatial Training Data -> GAN perform better than CNN in segmenting land cover classes outside of the training dataset (article, no code)
18.2. Building-A-Nets -> robust building extraction from high-resolution remote sensing images with adversarial networks
18.3. GANmapper -> a building footprint generator using Generative Adversarial Networks
18.4. CSA-CDGAN -> Channel Self-Attention Based Generative Adversarial Network for Change Detection of Remote Sensing Images
18.5. DSGAN -> a conditinal GAN for dynamic precipitation downscaling
18.6. MarsGAN -> GAN trained on satellite photos of Mars
18.7. HC_ADGAN -> codes for the paper Adaptive Dropblock Enhanced GenerativeAdversarial Networks for Hyperspectral Image Classification
18.8. SCALAE -> code for our paper Formatting the Landscape: Spatial conditional GAN for varying population in satellite imagery. Method to generate satellite imagery from custom 2D population maps
18.9. Satellite-Image-Forgery-Detection-and-Localization
18.10. STGAN -> PyTorch Implementation of STGAN for Cloud Removal in Satellite Images, with paper
18.11. ds-gan-spatiotemporal-evaluation -> evaluating use of deep generative models in remote sensing applications
18.12. pub-ffi-gan -> code for 2018 paper: Applying generative adversarial networks for anomaly detection in hyperspectral remote sensing imagery
18.14. Remote-Sensing-Image-Generation -> Generate RS Images using Generative Adversarial Networks (GAN)
18.15. RoadDA -> code for 2021 paper: Stagewise Unsupervised Domain Adaptation with Adversarial Self-Training for Road Segmentation of Remote Sensing Images
18.16. PSGan-Family -> code for 2020 paper: PSGAN: A Generative Adversarial Network for Remote Sensing Image Pan-Sharpening
18.17. Satellite Image Augmetation with GANs -> code for 2022 paper: Image Augmentation for Satellite Images
18.18. opt2sar-cyclegan -> Research on SAR image generation method based on non-homologous data
18.19. sentinel-cgan -> code for article: Generative adversarial networks in satellite image datasets augmentation
18.20. Shoreline_Extraction_GAN -> Shoreline extraction via generative adversarial networks, prediction via LSTMs
18.21. Landsat8-Sentinel2-Fusion -> Translating Landsat 8 to Sentinel-2 using a GAN
19. Autoencoders, dimensionality reduction, image embeddings & similarity search
Example of using an autoencoder to create a low dimensional representation of hyperspectral data.
Autoencoders are a type of neural network that aim to simplify the representation of input data by compressing it into a lower dimensional form. This is achieved through a two-step process of encoding and decoding, where the encoding step compresses the data into a lower dimensional representation, and the decoding step restores the data back to its original form. The goal of this process is to reduce the data's dimensionality, making it easier to store and process, while retaining the essential information. Dimensionality reduction, as the name suggests, refers to the process of reducing the number of dimensions in a dataset. This can be achieved through various techniques such as principal component analysis (PCA) or singular value decomposition (SVD). Autoencoders are one type of neural network that can be used for dimensionality reduction. In the field of computer vision, image embeddings are vector representations of images that capture the most important features of the image. These embeddings can then be used to perform similarity searches, where images are compared based on their features to find similar images. This process can be used in a variety of applications, such as image retrieval, where images are searched based on certain criteria like color, texture, or shape. It can also be used to identify duplicate images in a dataset. Image source
19.1. Autoencoders & their Application in Remote Sensing -> intro article and example use case applied to SAR data for land classification
19.2. LEt-SNE -> Dimensionality Reduction and visualization technique that compensates for the curse of dimensionality
19.3. AutoEncoders for Land Cover Classification of Hyperspectral Images -> An autoencoder nerual net is used to reduce 103 band data to 60 features (dimensionality reduction), keras. Also read part 2 which implements K-NNC, SVM and Gradient Boosting
19.4. Image-Similarity-Search -> an app that helps perform super fast image retrieval on PyTorch models for better embedding space interpretability
19.5. Interactive-TSNE -> a tool that provides a way to visually view a PyTorch model's feature representation for better embedding space interpretability
19.6. How Airbus Detects Anomalies in ISS Telemetry Data Using TFX -> uses an autoencoder
19.7. RoofNet -> identify roof age using historical satellite images to lower the customer acquisition cost for new solar installations. Uses a VAE: Variational Autoencoder
19.8. Visual search over billions of aerial and satellite images -> implemented at Descartes labs
19.9. parallax -> Tool for interactive embeddings visualization
19.10. Deep-Gapfill -> Official implementation of Optical image gap filling using deep convolutional autoencoder from optical and radar images
19.11. Mxnet repository for generating embeddings on satellite images -> Includes sampling of images, mining algorithms, different architectures, error functions, measures for evaluation.
19.12. Fine tuning CLIP with Remote Sensing (Satellite) images and captions -> fine tuning CLIP on the RSICD image captioning dataset, to enable querying large catalogues in natural language. With repo, uses
19.13. Image search with 🤗 datasets -> tutorial on fine tuning an image search model
19.14. SynImageAnalysis -> comparing synthetic and real satellite images in the latent feature space (embeddings)
19.15. GRN-SNDL -> model the relations between samples (or scenes) by making use of a graph structure which is fed into network learning
19.16. SauMoCo -> codes for TGRS paper: Deep Unsupervised Embedding for Remotely Sensed Images Based on Spatially Augmented Momentum Contrast
19.17. TGRS_RiDe -> Rotation Invariant Deep Embedding for RemoteSensing Images
19.18. RaVAEn -> RaVAEn is a lightweight, unsupervised approach for change detection in satellite data based on Variational Auto-Encoders (VAEs) with the specific purpose of on-board deployment
19.19. Reverse image search using deep discrete feature extraction and locality-sensitive hashing
19.20. SNCA_CE -> code for the paper Deep Metric Learning based on Scalable Neighborhood Components for Remote Sensing Scene Characterization
19.21. LandslideDetection-from-satellite-imagery -> Using Attention and Autoencoder boosted CNN
19.22. split-brain-remote-sensing -> code for 2018 paper: Analysis of Color Space Quantization in Split-Brain Autoencoder for Remote Sensing Image Classification
19.23. image-similarity-measures -> Implementation of eight evaluation metrics to access the similarity between two images. Blog post here
19.24. Large_Scale_GeoVisual_Search -> ResNet architecture on UC Merced Land Use Dataset with hamming distance for similarity based search
19.25. geobacter -> Generates useful feature embeddings for geospatial locations
19.26. Satellite-Image-Segmentation -> the KV-Net model uses this feature of autoencoders to reconnect the disconnected roads
19.27. Satellite-Image-Enhancement -> Image enhancement using GAN's and autoencoders
19.28. Variational-Autoencoder-For-Satellite-Imagery -> a special VAE to squeeze N images into one single representation with colors segmentating the different objects
19.29. DINCAE -> Data-Interpolating Convolutional Auto-Encoder is a neural network to reconstruct missing data in satellite observations
19.30. 3D_SITS_Clustering -> code for 2020 paper: Unsupervised Satellite Image Time Series Clustering Using Object-Based Approaches and 3D Convolutional Autoencoder
19.31. sat_cnn -> code for 2022 paper: Estimating Generalized Measures of Local Neighbourhood Context from Multispectral Satellite Images Using a Convolutional Neural Network. Uses a convolutional autoencoder (CAE)
19.32. you-are-here -> Matlab code for 2020 paper: You Are Here: Geolocation by Embedding Maps and Images
19.33. Tensorflow similarity -> offers state-of-the-art algorithms for metric learning and all the necessary components to research, train, evaluate, and serve similarity-based models
19.34. Train SimSiam on Satellite Images using lightly.ai to generate embeddings that can be used for data exploration and understanding
19.35. Airbus_SDC_dup -> Project focused on detecting duplicate regions of overlapping satellite imagery. Applied to Airbus ship detection dataset
19.36. scale-mae -> code for 2022 paper: Scale-MAE: A Scale-Aware Masked Autoencoder for Multiscale Geospatial Representation Learning
20. Image retrieval
Illustration of the remote sensing image retrieval process.
Image retrieval is the task of retrieving images from a collection that are similar to a query image. Image retrieval plays a vital role in remote sensing by enabling the efficient and effective search for relevant images from large image archives, and by providing a way to quantify changes in the environment over time. Image source
20.1. Demo_AHCL_for_TGRS2022 -> code for 2022 paper: Asymmetric Hash Code Learning (AHCL) for remote sensing image retrieval
20.2. GaLR -> code for 2022 paper: Remote Sensing Cross-Modal Text-Image Retrieval Based on Global and Local Information
20.3. retrievalSystem -> cross-modal image retrieval system
20.4. AMFMN -> code for the 2021 paper: Exploring a Fine-grained Multiscale Method for Cross-modal Remote Sensing Image Retrieval
20.5. Active-Learning-for-Remote-Sensing-Image-Retrieval -> unofficial implementation of paper: A Novel Active Learning Method in Relevance Feedback for Content-Based Remote Sensing Image Retrieval
20.6. CMIR-NET -> code for 2020 paper: A deep learning based model for cross-modal retrieval in remote sensing
20.7. Deep-Hash-learning-for-Remote-Sensing-Image-Retrieval -> code for 2020 paper: Deep Hash Learning for Remote Sensing Image Retrieval
20.8. MHCLN -> code for 2018 paper: Deep Metric and Hash-Code Learning for Content-Based Retrieval of Remote Sensing Images
20.9. HydroViet_VOR -> Object Retrieval in satellite images with Triplet Network
20.10. AMFMN -> code for 2021 paper: Exploring a Fine-Grained Multiscale Method for Cross-Modal Remote Sensing Image Retrieval
21. Image Captioning
Image Captioning is the task of automatically generating a textual description of an image. In remote sensing, image captioning can be used to automatically generate captions for satellite or aerial images, which can be useful for a variety of purposes, such as image search and retrieval, data cataloging, and data dissemination. The generated captions can provide valuable information about the content of the images, including the location, the type of terrain or objects present, and the weather conditions, among others. This information can be used to quickly and easily understand the content of the images, without having to manually examine each image. Image source
21.1. awesome-remote-image-captioning -> a list of awesome remote sensing image captioning resources
21.2. Fine tuning CLIP with Remote Sensing (Satellite) images and captions -> fine tuning CLIP on the RSICD image captioning dataset, to enable querying large catalogues in natural language. With repo, uses
21.3. CapFormer -> Pure transformer for remote sensing image caption
21.4. remote_sensing_image_captioning -> code for 2019 paper: Region Driven Remote Sensing Image Captioning
21.5. Remote Sensing Image Captioning with Transformer and Multilabel Classification
21.6. Siamese-spatial-Graph-Convolution-Network -> code for 2019 paper: Siamese graph convolutional network for content based remote sensing image retrieval
21.7. MLAT -> code for 2022 paper: Remote-Sensing Image Captioning Based on Multilayer Aggregated Transformer
21.8. WordSent -> code for 2020 paper: Word–Sentence Framework for Remote Sensing Image Captioning
21.9. a-mask-guided-transformer-with-topic-token -> code for 2022 paper: A Mask-Guided Transformer Network with Topic Token for Remote Sensing Image Captioning
21.10. MetaCaptioning -> code for 2022 paper: Meta captioning: A meta learning based remote sensing image captioning framework
21.11. Transformer-for-image-captioning -> a transformer for image captioning, trained on the UCM dataset
21.12. remote-sensing-image-caption -> image classification and image caption by PyTorch
22. Visual Question Answering
Visual Question Answering (VQA) is the task of automatically answering a natural language question about an image. In remote sensing, VQA enables users to interact with the images and retrieve information using natural language questions. For example, a user could ask a VQA system questions such as "What is the type of land cover in this area?", "What is the dominant crop in this region?" or "What is the size of the city in this image?". The system would then analyze the image and generate an answer based on its understanding of the image content.
22.1. VQA-easy2hard -> code for 2022 paper: From Easy to Hard: Learning Language-guided Curriculum for Visual Question Answering on Remote Sensing Data
22.2. lit4rsvqa -> code for paper: LiT-4-RSVQA: Lightweight Transformer-based Visual Question Answering in Remote Sensing
23. Mixed data learning
Mixed data learning is the process of learning from datasets that may contain an mix of images, textual and numeric data. Mixed data learning can help improve the accuracy of models by allowing them to learn from multiple sources at once and use more sophisticated methods to identify patterns and correlations.
23.1. Predicting the locations of traffic accidents with satellite imagery and convolutional neural networks -> Combining satellite imagery and structured data to predict the location of traffic accidents with a neural network of neural networks, with repo
23.2. Multi-Input Deep Neural Networks with PyTorch-Lightning - Combine Image and Tabular Data -> excellent intro article using pytorch, not actually applied to satellite data but to real estate data, with repo
23.3. Joint Learning from Earth Observation and OpenStreetMap Data to Get Faster Better Semantic Maps -> fusion based architectures and coarse-to-fine segmentation to include the OpenStreetMap layer into multispectral-based deep fully convolutional networks, arxiv paper
23.4. Composing Decision Forest and Neural Network models tensorflow documentation
23.5. pyimagesearch article on mixed-data
23.6. pytorch-widedeep -> A flexible package for multimodal-deep-learning to combine tabular data with text and images using Wide and Deep models in Pytorch
23.7. accidentRiskMap -> Inferring high-resolution traffic accident risk maps based on satellite imagery and GPS trajectories
23.8. Sub-meter resolution canopy height map by Meta -> Satellite Metadata combined with outputs from simple CNN to regress canopy height
24. Few-shot learning
This is a class of techniques which attempt to make predictions for classes with few, one or even zero examples provided during training. In zero shot learning (ZSL) the model is assisted by the provision of auxiliary information which typically consists of descriptions/semantic attributes/word embeddings for both the seen and unseen classes at train time (ref). These approaches are particularly relevant to remote sensing, where there may be many examples of common classes, but few or even zero examples for other classes of interest.
24.2. FSODM -> Official Code for paper "Few-shot Object Detection on Remote Sensing Images" on arxiv
24.3. Few-Shot Classification of Aerial Scene Images via Meta-Learning -> 2020 publication, a classification model that can quickly adapt to unseen categories using only a few labeled samples
24.4. Papers about Few-shot Learning / Meta-Learning on Remote Sensing
24.5. SPNet -> code for 2021 paper: Siamese-Prototype Network for Few-Shot Remote Sensing Image Scene Classification
24.6. MDL4OW -> code for 2020 paper: Few-Shot Hyperspectral Image Classification With Unknown Classes Using Multitask Deep Learning
24.7. P-CNN -> code for 2021 paper: Prototype-CNN for Few-Shot Object Detection in Remote Sensing Images
24.8. CIR-FSD-2022 -> code for 2022 paper: Context Information Refinement for Few-Shot Object Detection in Remote Sensing Images
24.9. IEEE_TNNLS_Gia-CFSL -> code for 2022 paper: Graph Information Aggregation Cross-Domain Few-Shot Learning for Hyperspectral Image Classification
24.10. TIP_2022_CMFSL -> code for 2022 paper: Few-shot Learning with Class-Covariance Metric for Hyperspectral Image Classification
24.11. sen12ms-human-few-shot-classifier -> code for paper: Humans are poor few-shot classifiers for Sentinel-2 land cover
24.12. S3Net -> code for 2022 paper: S3Net: Spectral–Spatial Siamese Network for Few-Shot Hyperspectral Image Classification
24.13. SiameseNet-for-few-shot-Hyperspectral-Classification -> code for 2020 paper: 3DCSN:SiameseNet-for-few-shot-Hyperspectral-Classification
25. Self-supervised, unsupervised & contrastive learning
Self-supervised, unsupervised & contrastive learning are all methods of machine learning that use unlabeled data to train algorithms. Self-supervised learning uses labeled data to create an artificial supervisor, while unsupervised learning uses only the data itself to identify patterns and similarities. Contrastive learning uses pairs of data points to learn representations of data, usually for classification tasks.
25.1. Seasonal Contrast: Unsupervised Pre-Training from Uncurated Remote Sensing Data -> Seasonal Contrast (SeCo) is an effective pipeline to leverage unlabeled data for in-domain pre-training of remote sensing representations. Models trained with SeCo achieve better performance than their ImageNet pre-trained counterparts and state-of-the-art self-supervised learning methods on multiple downstream tasks. paper and repo
25.2. Unsupervised Learning for Land Cover Classification in Satellite Imagery
25.3. Tile2Vec: Unsupervised representation learning for spatially distributed data
25.4. Contrastive Sensor Fusion -> Code implementing Contrastive Sensor Fusion, an approach for unsupervised learning of multi-sensor representations targeted at remote sensing imagery
25.5. hyperspectral-autoencoders -> Tools for training and using unsupervised autoencoders and supervised deep learning classifiers for hyperspectral data, built on tensorflow. Autoencoders are unsupervised neural networks that are useful for a range of applications such as unsupervised feature learning and dimensionality reduction.
25.6. Sentinel-2 image clustering in python
25.7. MARTA GANs: Unsupervised Representation Learning for Remote Sensing Image Classification and code
25.8. A generalizable and accessible approach to machine learning with global satellite imagery nature publication -> MOSAIKS is designed to solve an unlimited number of tasks at planet-scale quickly using feature vectors, with repo. Also see mosaiks-api
25.9. contrastive-satellite -> Using contrastive learning to create embeddings from optical EuroSAT Satellite-2 imagery
25.10. Self-Supervised Learning of Remote Sensing Scene Representations Using Contrastive Multiview Coding -> arxiv paper and code
25.11. Self-Supervised-Learner by spaceml-org -> train a classifier with fewer labeled examples needed using self-supervised learning, example applied to UC Merced land use dataset
25.12. deepsentinel -> a sentinel-1 and -2 self-supervised sensor fusion model for general purpose semantic embedding
25.13. contrastive_SSL_ship_detection -> Contrastive self supervised learning for ship detection in Sentinel 2 images
25.14. geography-aware-ssl -> uses spatially aligned images over time to construct temporal positive pairs in contrastive learning and geo-location to design pre-text tasks
25.15. CNN-Supervised Classification -> Python code for self-supervised classification of remotely sensed imagery - part of the Deep Riverscapes project
25.16. clustimage -> a python package for unsupervised clustering of images
25.17. LandSurfaceClustering -> Land surface classification using remote sensing data with unsupervised machine learning (k-means)
25.18. K-Means Clustering for Surface Segmentation of Satellite Images
25.19. Sentinel-2 satellite imagery for crop classification using unsupervised clustering -> label groups of pixels based on temporal trends of their NDVI values
25.20. TheColorOutOfSpace -> Pytorch code for the paper "The color out of space: learning self-supervised representations for Earth Observation imagery" using the BigEarthNet dataset
25.21. Semantic segmentation of SAR images using a self supervised technique
25.22. STEGO -> Unsupervised Semantic Segmentation by Distilling Feature Correspondences, with paper
25.23. Unsupervised Segmentation of Hyperspectral Remote Sensing Images with Superpixels
25.24. SoundingEarth -> Self-supervised Audiovisual Representation Learning for Remote Sensing Data, uses the SoundingEarth Dataset
25.25. singleSceneSemSegTgrs2022 -> code for 2022 paper: Unsupervised Single-Scene Semantic Segmentation for Earth Observation
25.26. SSLRemoteSensing -> code for 2021 paper: Semantic Segmentation of Remote Sensing Images With Self-Supervised Multitask Representation Learning
25.27. CBT code for 2022 paper: Continual Barlow Twins: continual self-supervised learning for remote sensing semantic segmentation
25.28. Unsupervised Satellite Image Classification based on Partial Adversarial Domain Adaptation -> Code for course project
25.29. T2FTS -> code for 2022 paper: Teaching Teachers First and Then Student: Hierarchical Distillation to Improve Long-Tailed Object Recognition in Aerial Images
25.30. SSLTransformerRS -> code for 2022 paper: Self-supervised Vision Transformers for Land-cover Segmentation and Classification
25.31. DINO-MM -> code for 2022 paper: Self-supervised Vision Transformers for Joint SAR-optical Representation Learning
25.32. SSL4EO-S12 -> a large-scale dataset for self-supervised learning in Earth observation
25.33. SSL4EO-Review -> code for 2022 paper: Self-supervised Learning in Remote Sensing: A Review
25.34. transfer_learning_cspt -> code for 2022 paper: Consecutive Pretraining: A Knowledge Transfer Learning Strategy with Relevant Unlabeled Data for Remote Sensing Domain
25.35. OTL -> code for 2022 paper: Clustering-Based Representation Learning through Output Translation and Its Application to Remote-Sensing Images
25.36. Push-and-Pull-Network -> code for 2022 paper: Contrastive Learning for Fine-grained Ship Classification in Remote Sensing Images
25.37. vissl_experiments -> Self-supervised Learning using Facebook VISSL on the RESISC-45 satellite imagery classification dataset
25.38. MS2A-Net -> code for 2022 paper: MS 2 A-Net: Multi-scale spectral-spatial association network for hyperspectral image clustering
25.39. UDA_for_RS -> code for paper: Unsupervised Domain Adaptation for Remote Sensing Semantic Segmentation with Transformer
25.40. pytorch-ssl-building_extract -> code for 2022 paper: Research on Self-Supervised Building Information Extraction with High-Resolution Remote Sensing Images for Photovoltaic Potential Evaluation
25.41. self-rare-wildlife -> code for 2021 paper: Self-Supervised Pretraining and Controlled Augmentation Improve Rare Wildlife Recognition in UAV Images
25.42. SatMAE -> code for 2022 paper: SatMAE: Pre-training Transformers for Temporal and Multi-Spectral Satellite Imagery
25.43. FireCLR-Wildfires -> code for 2022 paper: Unsupervised Wildfire Change Detection based on Contrastive Learning
25.44. FALSE -> code for 2022 paper: False: False Negative Samples Aware Contrastive Learning for Semantic Segmentation of High-Resolution Remote Sensing Image
25.45. MATTER -> code for 2022 paper: Self-Supervised Material and Texture Representation Learning for Remote Sensing Tasks
26. Weakly & semi-supervised learning
Weakly & semi-supervised learning are two methods of machine learning that use both labeled and unlabeled data for training. Weakly supervised learning uses weakly labeled data, which may be incomplete or inaccurate, while semi-supervised learning uses both labeled and unlabeled data. Weakly supervised learning is typically used in situations where labeled data is scarce and unlabeled data is abundant. Semi-supervised learning is typically used in situations where labeled data is abundant but also contains some noise or errors. Both techniques can be used to improve the accuracy of machine learning models by making use of additional data sources.
26.1. MARE -> self-supervised Multi-Attention REsu-net for semantic segmentation in remote sensing
26.2. SSGF-for-HRRS-scene-classification -> code for 2018 paper: A semi-supervised generative framework with deep learning features for high-resolution remote sensing image scene classification
26.3. SFGAN -> code for 2018 paper: Semantic-Fusion Gans for Semi-Supervised Satellite Image Classification
26.4. SSDAN -> code for 2021 paper: Multi-Source Semi-Supervised Domain Adaptation Network for Remote Sensing Scene Classification
26.5. HR-S2DML -> code for 2020 paper: High-Rankness Regularized Semi-Supervised Deep Metric Learning for Remote Sensing Imagery
26.6. Semantic Segmentation of Satellite Images Using Point Supervision
26.7. fcd -> code for 2021 paper: Fixed-Point GAN for Cloud Detection. A weakly-supervised approach, training with only image-level labels
26.8. weak-segmentation -> Weakly supervised semantic segmentation for aerial images in pytorch
26.9. TNNLS_2022_X-GPN -> Code for paper: Semisupervised Cross-scale Graph Prototypical Network for Hyperspectral Image Classification
26.10. weakly_supervised -> code for the paper Weakly Supervised Deep Learning for Segmentation of Remote Sensing Imagery. Demonstrates that segmentation can be performed using small datasets comprised of pixel or image labels
26.11. wan -> Weakly-Supervised Domain Adaptation for Built-up Region Segmentation in Aerial and Satellite Imagery, with arxiv paper
26.12. sourcerer -> A Bayesian-inspired deep learning method for semi-supervised domain adaptation designed for land cover mapping from satellite image time series (SITS). Paper
26.13. MSMatch -> Semi-Supervised Multispectral Scene Classification with Few Labels. Includes code to work with both the RGB and the multispectral (MS) versions of EuroSAT dataset and the UC Merced Land Use (UCM) dataset. Paper
26.14. Flood Segmentation on Sentinel-1 SAR Imagery with Semi-Supervised Learning with arxiv paper
26.15. Semi-supervised learning in satellite image classification -> experimenting with MixMatch and the EuroSAT data set
26.16. ScRoadExtractor -> code for 2020 paper: Scribble-based Weakly Supervised Deep Learning for Road Surface Extraction from Remote Sensing Images
26.17. ICSS -> code for 2022 paper: Weakly-supervised continual learning for class-incremental segmentation
26.18. es-CP -> code for 2022 paper: Semi-Supervised Hyperspectral Image Classification Using a Probabilistic Pseudo-Label Generation Framework
26.19. Flood_Mapping_SSL -> code for 2022 paper: Enhancement of Urban Floodwater Mapping From Aerial Imagery With Dense Shadows via Semisupervised Learning
26.20. MS4D-Net-Building-Damage-Assessment -> code for 2022 paper: MS4D-Net: Multitask-Based Semi-Supervised Semantic Segmentation Framework with Perturbed Dual Mean Teachers for Building Damage Assessment from High-Resolution Remote Sensing Imagery
27. Active learning
Supervised deep learning techniques typically require a huge number of annotated/labelled examples to provide a training dataset. However labelling at scale take significant time, expertise and resources. Active learning techniques aim to reduce the total amount of annotation that needs to be performed by selecting the most useful images to label from a large pool of unlabelled images, thus reducing the time to generate useful training datasets. These processes may be referred to as Human-in-the-Loop Machine Learning
27.1. Active learning for object detection in high-resolution satellite images -> arxiv paper
27.2. AIDE V2 - Tools for detecting wildlife in aerial images using active learning
27.3. AstronomicAL -> An interactive dashboard for visualisation, integration and classification of data using Active Learning
27.4. Follow tutorials for active learning for object detection and segmentation on the lightly platform.
27.5. Active-Labeler by spaceml-org -> a CLI Tool that facilitates labeling datasets with just a SINGLE line of code
27.6. Labelling platform for Mapping Africa active learning project
27.7. ChangeDetectionProject -> Trying out Active Learning in with deep CNNs for Change detection on remote sensing data
27.8. ALS4GAN -> Active Learning for Improved Semi Supervised Semantic Segmentation in Satellite Images, with paper
27.9. Active-Learning-for-Remote-Sensing-Image-Retrieval -> unofficial implementation of paper: A Novel Active Learning Method in Relevance Feedback for Content-Based Remote Sensing Image Retrieval
27.10. DIAL -> code for 2022 paper: DIAL: Deep Interactive and Active Learning for Semantic Segmentation in Remote Sensing
27.11. whales -> An active learning pipeline for identifying whales in high-resolution satellite imagery, by Microsoft
28. Federated learning
Federated learning is an approach to distributed machine learning where a central processor coordinates the training of an individual model in each of its clients. It is a type of distributed ML which means that the data is distributed among different devices or locations and the model is trained on all of them. The central processor aggregates the model updates from all the clients and then sends the global model parameters back to the clients. This is done to protect the privacy of data, as the data remains on the local device and only the global model parameters are shared with the central processor. This technique can be used to train models with large datasets that cannot be stored in a single device, as well as to enable certain privacy-preserving applications.
28.1. Federated-Learning-for-Remote-Sensing -> implementation of three Federated Learning models
28.2. Semantic-Segmentation-UNet-Federated -> code for paper: FedUKD: Federated UNet Model with Knowledge Distillation for Land Use Classification from Satellite and Street Views
28.3. MM-FL -> code for paper: Learning Across Decentralized Multi-Modal Remote Sensing Archives with Federated Learning
29. Transformers
Vision transformers are state-of-the-art models for vision tasks such as image classification and object detection. They differ from CNNs as they use self-attention instead of convolution to learn global relations between all pixels in the image. Vision transformers employ a transformer encoder architecture, composed of multi-layer blocks with multi-head self-attention and feed-forward layers, enabling the capture of rich contextual information for more accurate predictions.
29.1. Transformer-in-Remote-Sensing -> code for 2022 paper: Transformers in Remote Sensing: A Survey
29.2. Remote-Sensing-RVSA -> code for 2022 paper: Advancing Plain Vision Transformer Towards Remote Sensing Foundation Model
29.3. SatViT -> self-supervised training of multispectral optical and SAR vision transformers
29.4. UDA_for_RS -> code for 2022 paper: Unsupervised Domain Adaptation for Remote Sensing Semantic Segmentation with Transformer
29.5. Vision Transformers for Low Earth Orbit Satellites -> blog post that investigates deploying Vision Transformers on low earth orbit satellites
30. Adversarial ML
Efforts to detect falsified images & deepfakes
30.1. UAE-RS -> dataset that provides black-box adversarial samples in the remote sensing field
30.2. PSGAN -> code for paper: Perturbation Seeking Generative Adversarial Networks: A Defense Framework for Remote Sensing Image Scene Classification
30.3. SACNet -> code for 2021 paper: Self-Attention Context Network: Addressing the Threat of Adversarial Attacks for Hyperspectral Image Classification
31. Image registration
Image registration is the process of registering one or more images onto another (typically well georeferenced) image. Traditionally this is performed manually by identifying control points (tie-points) in the images, for example using QGIS. This section lists approaches which mostly aim to automate this manual process. There is some overlap with the data fusion section but the distinction I make is that image registration is performed as a prerequisite to downstream processes which will use the registered data as an input.
31.1. Wikipedia article on registration -> register for change detection or image stitching
31.2. Phase correlation is used to estimate the XY translation between two images with sub-pixel accuracy. Can be used for accurate registration of low resolution imagery onto high resolution imagery, or to register a sub-image on a full image -> Unlike many spatial-domain algorithms, the phase correlation method is resilient to noise, occlusions, and other defects. With additional pre-processing image rotation and scale changes can also be calculated.
31.3. How to Co-Register Temporal Stacks of Satellite Images
31.4. ImageRegistration -> Interview assignment for multimodal image registration using SIFT
31.5. imreg_dft -> Image registration using discrete Fourier transform. Given two images it can calculate the difference between scale, rotation and position of imaged features. Used by the up42 co-registration service
31.6. arosics -> Perform automatic subpixel co-registration of two satellite image datasets using phase-correlation, XY translations only.
31.7. SubpixelAlignment -> Implementation of tiff image alignment through phase correlation for pixel- and subpixel-bias
31.8. cnn-registration -> A image registration method using convolutional neural network features written in Python2, Tensorflow 1.5
31.9. Detecting Ground Control Points via Convolutional Neural Network for Stereo Matching -> code?
31.10. ImageCoregistration -> Image registration with openCV using sift and RANSAC
31.11. mapalignment -> Aligning and Updating Cadaster Maps with Remote Sensing Images
31.12. CVPR21-Deep-Lucas-Kanade-Homography -> deep learning pipeline to accurately align challenging multimodality images. The method is based on traditional Lucas-Kanade algorithm with feature maps extracted by deep neural networks.
31.13. eolearn implements phase correlation, feature matching and ECC
31.14. RStoolbox supports Image to Image Co-Registration based on Mutual Information
31.15. Reprojecting the Perseverance landing footage onto satellite imagery
31.16. Kornia provides image registration
31.17. LoFTR -> Detector-Free Local Feature Matching with Transformers. Good performance matching satellite image pairs, tryout the web demo on your data
31.18. image-to-db-registration -> This remote module implements an algorithm for automated vector Database registration onto an Image. Implemented in the orfeo-toolbox
31.19. MS_HLMO_registration -> Multi-scale Histogram of Local Main Orientation for Remote Sensing Image Registration, with paper
31.20. cnn-matching -> code and datadset for paper: Deep learning algorithm for feature matching of cross modality remote sensing images
31.21. Imatch-P -> A demo using SuperGlue and SuperPoint to do the image matching task based PaddlePaddle
31.22. NBR-Net -> A Non-rigid Bi-directional Registration Network for Multi-temporal Remote Sensing Images
31.23. MU-Net -> code for paper: A Multi-Scale Framework with Unsupervised Learning for Remote Sensing Image Registration
31.24. unsupervisedDeepHomographyRAL2018 -> Unsupervised Deep Homography applied to aerial data
31.25. registration_cnn_ntg -> code for paper: A Multispectral Image Registration Method Based on Unsupervised Learning
31.26. remote-sensing-images-registration-dataset -> at 0.23m, 3.75m & 30m resolution
31.27. semantic-template-matching -> code for 2021 paper: A deep learning semantic template matching framework for remote sensing image registration
31.28. GMN-Generative-Matching-Network -> code for 2018 paper: Deep Generative Matching Network for Optical and SAR Image Registration
31.29. SOMatch -> code for 2020 paper: A deep learning framework for matching of SAR and optical imagery
31.30. Interspectral image registration dataset -> including satellite and drone imagery
31.31. RISG-image-matching -> A rotation invariant SuperGlue image matching algorithm
31.32. DeepAerialMatching_pytorch -> code for 2020 paper: A Two-Stream Symmetric Network with Bidirectional Ensemble for Aerial Image Matching
31.33. DPCN -> code for 2020 paper: Deep Phase Correlation for End-to-End Heterogeneous Sensor Measurements Matching
31.34. FSRA -> code for 2022 paper: A Transformer-Based Feature Segmentation and Region Alignment Method For UAV-View Geo-Localization
31.35. IHN -> code for 2022 paper: Iterative Deep Homography Estimation
31.36. OSMNet -> code for 2021 paper: Explore Better Network Framework for High-Resolution Optical and SAR Image Matching
31.37. L2_Siamese -> code for the 2020 paper: Registration of Multiresolution Remote Sensing Images Based on L2-Siamese Model
32. Terrain mapping, Disparity Estimation, Lidar, DEMs & NeRF
Measure surface contours & locate 3D points in space from 2D images. NeRF stands for Neural Radiance Fields and is the term used in deep learning communities to describe a model that generates views of complex 3D scenes based on a partial set of 2D images
32.1. Wikipedia DEM article and phase correlation article
32.2. Intro to depth from stereo
32.3. Map terrain from stereo images to produce a digital elevation model (DEM) -> high resolution & paired images required, typically 0.3 m, e.g. Worldview
32.4. Process of creating a DEM here
32.5. ArcGIS can generate DEMs from stereo images
32.6. https://github.com/MISS3D/s2p -> produces elevation models from images taken by high resolution optical satellites -> demo code on https://gfacciol.github.io/IS18/
32.7. Predict the fate of glaciers
32.8. monodepth - Unsupervised single image depth prediction with CNNs
32.9. Stereo Matching by Training a Convolutional Neural Network to Compare Image Patches
32.11. Phase correlation in scikit-image
32.12. s2p -> a Python library and command line tool that implements a stereo pipeline which produces elevation models from images taken by high resolution optical satellites such as Pléiades, WorldView, QuickBird, Spot or Ikonos
32.13. The Mapbox API provides images and elevation maps, article here
32.14. Reconstructing 3D buildings from aerial LiDAR with Mask R-CNN
32.15. ResDepth -> A Deep Prior For 3D Reconstruction From High-resolution Satellite Images
32.16. overhead-geopose-challenge -> competition to build computer vision algorithms that can effectively model the height and pose of ground objects for monocular satellite images taken from oblique angles. Blog post MEET THE WINNERS OF THE OVERHEAD GEOPOSE CHALLENGE
32.17. cars -> a dedicated and open source 3D tool to produce Digital Surface Models from satellite imaging by photogrammetry. This Multiview stereo pipeline is intended for massive DSM production with a robust and performant design
32.18. ImageToDEM -> Generating Elevation Surface from a Single RGB Remotely Sensed Image Using a U-Net for generator and a PatchGAN for the discriminator
32.19. IMELE -> Building Height Estimation from Single-View Aerial Imagery
32.20. ridges -> deep semantic segmentation model for identifying ridges in topography
32.21. planet_tools -> Selection of imagery from Planet API for creation of stereo elevation models
32.22. SatelliteNeRF -> PyTorch-based Neural Radiance Fields adapted to satellite domain
32.23. SatelliteSfM -> A library for solving the satellite structure from motion problem
32.24. SatelliteSurfaceReconstruction -> 3D Surface Reconstruction From Multi-Date Satellite Images, ISPRS, 2021
32.25. son2sat -> A neural network coded in TensorFlow 1 that produces satellite images from acoustic images
32.26. aerial_mtl -> PyTorch implementation for multi-task learning with aerial images to learn both semantics and height from aerial image datasets; fuses RGB & lidar
32.27. ReKlaSat-3D -> 3D Reconstruction and Classification from Very High Resolution Satellite Imagery
32.28. M3Net -> A deep learning method for building height estimation using high-resolution multi-view imagery over urban areas
32.29. HMSM-Net -> code for 2022 paper: Hierarchical multi-scale matching network for disparity estimation of high-resolution satellite stereo images
32.30. StereoMatchingRemoteSensing -> code for 2021 paper: Dual-Scale Matching Network for Disparity Estimation of High-Resolution Remote Sensing Images
32.31. satnerf -> Learning Multi-View Satellite Photogrammetry With Transient Objects and Shadow Modeling Using RPC Cameras
32.32. SatMVS -> code for 2021 paper: Rational Polynomial Camera Model Warping for Deep Learning Based Satellite Multi-View Stereo Matching
32.33. ImpliCity -> reconstructs digital surface models (DSMs) from raw photogrammetric 3D point clouds and ortho-images with the help of an implicit neural 3D scene representation
32.34. WHU-Stereo -> a large-scale dataset for stereo matching of high-resolution satellite imagery & several deep learning methods for stereo matching. Methods include StereoNet, Pyramid Stereo Matching Network & HMSM-Net
32.35. Photogrammetry-Guide -> A guide covering Photogrammetry including the applications, libraries and tools that will make you a better and more efficient Photogrammetry development
32.36. DSM-to-DTM -> Exploring the use of machine learning to convert a Digital Surface Model (e.g. SRTM) to a Digital Terrain Model
32.37. GF-7_Stereo_Matching -> code for paper: Large Scene DSM Generation of Gaofen-7 Imagery Combined with Deep Learning
32.38. Mapping drainage ditches in forested landscapes using deep learning and aerial laser scanning
32.39. 3DCD -> code for paper: Inferring 3D change detection from bitemporal optical images
33. Thermal Infrared
Thermal infrared remote sensing is a technique used to detect and measure thermal radiation emitted from the Earth’s surface. This technique can be used to measure the temperature of the ground and any objects on it and can detect the presence of different materials. Thermal infrared remote sensing is used to assess land cover, detect land-use changes, and monitor urban heat islands, as well as to measure the temperature of the ground during nighttime or in areas of limited visibility.
33.1. The World Needs (a lot) More Thermal Infrared Data from Space
33.2. IR2VI thermal-to-visible image translation framework based on GANs with code
33.3. Object_Classification_in_Thermal_Images -> classification accuracy was improved by adding the object size as a feature directly within the CNN
33.4. Thermal imaging with satellites blog post by Christoph Rieke
34. SAR
SAR (synthetic aperture radar) is used to detect and measure the properties of objects and surfaces on the Earth's surface. SAR can be used to detect changes in terrain, features, and objects over time, as well as to measure the size, shape, and composition of objects and surfaces. SAR can also be used to measure moisture levels in soil and vegetation, or to detect and monitor changes in land use.
34.1. awesome-sar -> A curated list of awesome Synthetic Aperture Radar (SAR) software, libraries, and resources
34.2. Removing speckle noise from Sentinel-1 SAR using a CNN
34.3. You do not need clean images for SAR despeckling with deep learning -> How Speckle2Void learned to stop worrying and love the noise
34.4. PySAR - InSAR (Interferometric Synthetic Aperture Radar) timeseries analysis in python
34.5. Synthetic Aperture Radar (SAR) Analysis With Clarifai
34.6. Labeled SAR imagery dataset of ten geophysical phenomena from Sentinel-1 wave mode consists of more than 37,000 SAR vignettes divided into ten defined geophysical categories
34.7. Deep Learning and SAR Applications
34.9. s1_parking_occupancy -> Source code for PARKING OCCUPANCY ESTIMATION ON SENTINEL-1 IMAGES, ISPRS 2020
34.11. SpaceNet_SAR_Buildings_Solutions -> The winning solutions for the SpaceNet 6 Challenge
34.12. Mapping and monitoring of infrastructure in desert regions with Sentinel-1
34.13. xView3 is a competition to detect dark vessels using computer vision and global SAR satellite imagery. First place solution and second place solution. Additional places up to fifth place are available at the xView GitHub Organization page
34.14. Winners of the STAC Overflow: Map Floodwater from Radar Imagery competition
34.15. deSpeckNet-TF-GEE -> implementation of the paper 'deSpeckNet: Generalizing Deep Learning Based SAR Image Despeckling'
34.16. cnn_sar_image_classification -> CNN for classifying SAR images of the Amazon Rainforest
34.17. s1_icetype_cnn -> Retrieve sea ice type from Sentinel-1 SAR with CNN
34.18. SARSeg -> pytorch code for the paper 'MP-ResNet: Multi-path Residual Network for the Semantic segmentation of PolSAR Images'
34.19. TGRS_DisOptNet -> Distilling Semantic Knowledge from Optical Images for Weather-independent Building Segmentation
34.20. SAR_CD_DDNet -> PyTorch implementation of Change Detection in Synthetic Aperture Radar Images Using a Dual Domain Network
34.21. SAR_CD_MS_CapsNet -> Code for the paper "Change Detection in SAR Images Based on Multiscale Capsule Network" IEEE GRSL 2021
34.22. anomaly-detection-in-SAR-imagery -> identify an unknown ship in docks using keras & retinanet
34.23. sar_transformer -> Transformer based SAR image despeckling, trained with synthetic imagery, with paper
34.24. SSDD ship detection dataset
34.25. Semantic segmentation of SAR images using a self supervised technique
34.26. Ship Detection on Remote Sensing Synthetic Aperture Radar Data -> based on the architectures of the Faster-RCNN and YOLOv5 networks
34.27. Target Recognition in SAR -> Identify Military Vehicles in Satellite Imagery with TensorFlow, with article
34.28. DSN -> code for 2020 paper: Deep SAR-Net: Learning objects from signals
34.29. SAR_denoising -> project on application of FFDNet to SAR images
34.30. cnninsar -> code for 2018 paper: CNN-Based InSAR Denoising and Coherence Metric
34.31. sar -> Despeckling Synthetic Aperture Radar Images using a Deep Residual CNN
34.32. GCBANet -> code for 2022 paper: A Global Context Boundary-Aware Network for SAR Ship Instance Segmentation
34.33. SAR_CD_GKSNet -> code for 2022 paper: Change Detection from Synthetic Aperture Radar Images via Graph-Based Knowledge Supplement Network
34.34. pixel-wise-segmentation-of-sar -> code for 2020 paper: Pixel-Wise Segmentation of SAR Imagery Using Encoder-Decoder Network and Fully-Connected CRF
34.35. SAR_Ship_detection_CFAR -> An improved two-parameter CFAR algorithm based on Rayleigh distribution and Mathematical Morphology for SAR ship detection
34.36. sar_snow_melt_timing -> notebooks and tools to identify snowmelt timing using timeseries analysis of backscatter of Sentinel-1 C-band SAR
34.37. Denoising radar satellite images using deep learning in Python -> Medium article on deepdespeckling
34.38. random-wetlands -> Random forest classification for wetland vegetation from synthetic aperture radar dataset
34.39. AGSDNet -> code for 2022 paper: AGSDNet: Attention and Gradient-Based SAR Denoising Network
34.40. LFG-Net -> code for 2022 paper: LFG-Net: Low-Level Feature Guided Network for Precise Ship Instance Segmentation in SAR Images
34.41. sar_sift -> Image registration algorithm
34.42. SAR-Despeckling -> toolbox
34.43. cogsima2022 -> code for 2022 paper: Enhancing land subsidence awareness via InSAR data and Deep Transformers
34.44. XAI4SAR-PGIL -> code for 2021 paper: Physically Explainable CNN for SAR Image Classification
34.45. PolSARFormer -> code for 2023 paper: Local Window Attention Transformer for Polarimetric SAR Image Classification
34.46 Toushka Waterbodies Segmentation from four different combinations of Sentinel-1 SAR imagery and Digital Elevation Model with Pytorch and U-net. -> code
35. NDVI - vegetation index
Normalized Difference Vegetation Index (NDVI) is an index used to measure the amount of healthy vegetation in a given area. It is calculated by taking the difference between the near-infrared (NIR) and red (red) bands of a satellite image, and dividing by the sum of the two bands. NDVI can be used to identify areas of healthy vegetation and to assess the health of vegetation in a given area.
35.1. Calculated via band math ndvi = np.true_divide((ir - r), (ir + r))
but challenging due to the size of the imagery
35.2. Example notebook local
35.3. Landsat data in cloud optimised (COG) format analysed for NDVI with medium article here.
35.4. Identifying Buildings in Satellite Images with Machine Learning and Quilt -> NDVI & edge detection via gaussian blur as features, fed to TPOT for training with labels from OpenStreetMap, modelled as a two class problem, “Buildings” and “Nature”
35.5. Seeing Through the Clouds - Predicting Vegetation Indices Using SAR
35.6. A walkthrough on calculating NDWI water index for flooded areas -> Derive zonal statistics from Sentinel 2 images using Rasterio and Geopandas
35.7. NDVI-Net -> code for 2020 paper: NDVI-Net: A fusion network for generating high-resolution normalized difference vegetation index in remote sensing
35.8. Awesome-Vegetation-Index
35.9. Remote-Sensing-Indices-Derivation-Tool -> Calculate spectral remote sensing indices from satellite imagery
36. General image quality
Image quality describes the degree of accuracy with which an image can represent the original object. Image quality is typically measured by the amount of detail, sharpness, and contrast that an image contains. Factors that contribute to image quality include the resolution, format, and compression of the image.
36.1. Convolutional autoencoder network can be employed to image denoising, read about this on the Keras blog
36.2. jitter-compensation -> Remote Sensing Image Jitter Detection and Compensation Using CNN
36.3. DeblurGANv2 -> Deblurring (Orders-of-Magnitude) Faster and Better
36.4. image-quality-assessment -> CNN to predict the aesthetic and technical quality of images
36.5. Convolutional autoencoder for image denoising -> keras guide
36.6. piq -> a collection of measures and metrics for image quality assessment
36.7. FFA-Net -> Feature Fusion Attention Network for Single Image Dehazing
36.8. DeepCalib -> A Deep Learning Approach for Automatic Intrinsic Calibration of Wide Field-of-View Cameras
36.9. PerceptualSimilarity -> LPIPS is a perceptual metric which aims to overcome the limitations of traditional metrics such as PSNR & SSIM, to better represent the features the human eye picks up on
36.10. Optical-RemoteSensing-Image-Resolution -> code for 2018 paper: Deep Memory Connected Neural Network for Optical Remote Sensing Image Restoration. Two applications: Gaussian image denoising and single image super-resolution
36.11. Hyperspectral-Deblurring-and-Destriping
36.12. HyDe -> Hyperspectral Denoising algorithm toolbox in Python, with paper
36.13. HLF-DIP -> code for 2022 paper: Unsupervised Hyperspectral Denoising Based on Deep Image Prior and Least Favorable Distribution
36.14. RQUNetVAE -> code for 2022 paper: Riesz-Quincunx-UNet Variational Auto-Encoder for Satellite Image Denoising
36.15. deep-hs-prior -> code for 2019 paper: Deep Hyperspectral Prior: Denoising, Inpainting, Super-Resolution
36.16. iquaflow -> from Satellogic, an image quality framework that aims at providing a set of tools to assess image quality by using the performance of AI models trained on the images as a proxy.
36.17. lvrnet -> Lightweight Image Restoration for Aerial Images under Low Visibility
37. Synthetic data
Training data can be hard to acquire, particularly for rare events such as change detection after disasters, or imagery of rare classes of objects. In these situations, generating synthetic training data might be the only option. This has become quite sophisticated, with 3D models being use with open source games engines such as Unreal.
37.1. The Synthinel-1 dataset: a collection of high resolution synthetic overhead imagery for building segmentation with repo
37.2. RarePlanes -> incorporates both real and synthetically generated satellite imagery including aircraft. Read the arxiv paper and checkout this repo. Note the dataset is available through the AWS Open-Data Program for free download
37.3. Read this article from NVIDIA which discusses fine tuning a model pre-trained on synthetic data (Rareplanes) with 10% real data, then pruning the model to reduce its size, before quantizing the model to improve inference speed
37.4. Combining Synthetic Data with Real Data to Improve Detection Results in Satellite Imagery
37.5. BlenderGIS could be used for synthetic data generation
37.6. bifrost.ai -> simulated data service with geospatial output data formats
37.7. oktal-se -> software for generating simulated data across a wide range of bands including optical and SAR
37.8. The Nuances of Extracting Utility from Synthetic Data -> We find that strategically augmenting the real dataset is nearly as effective as adding synthetic data in the quest to improve the detection or rare object classes, and that fully extracting the utility of synthetic data is a nuanced process
37.9. Synthesizing Robustness -> explores how to best leverage and enhance synthetic data
37.10. rendered.ai -> The Platform as a Service for Creating Synthetic Data
37.11. synthetic_xview_airplanes -> creation of airplanes synthetic dataset using ArcGIS CityEngine
37.13. SynImageAnalysis -> comparing syn and real sattlelite images in the latent feature space (embeddings)
37.14. Import OpenStreetMap data into Unreal Engine 4
37.15. deepfake-satellite-images -> dataset that includes over 1M images of synthetic aerial images
37.16. synthetic-disaster -> Generate synthetic satellite images of natural disasters using deep neural networks
37.17. STPLS3D -> A Large-Scale Synthetic and Real Aerial Photogrammetry 3D Point Cloud Dataset
37.18. LESS -> LargE-Scale remote sensing data and image Simulation framework over heterogeneous 3D scenes
37.19. Synthesizing Robustness: Dataset Size Requirements and Geographic Insights -> Medium article, concludes that synthetic data is most beneficial to the rarest object classes and that extracting utility from synthetic data often takes significant effort and creativity
37.20. rs_img_synth -> code for 2020 paper: Synthesizing Optical and SAR Imagery From Land Cover Maps and Auxiliary Raster Data
37.21. OnlyPlanes -> dataset and pretrained models for the paper: OnlyPlanes - Incrementally Tuning Synthetic Training Datasets for Satellite Object Detection
- Logo created with Brandmark