Training sensor-agnostic deep learning models for remote sensing: Achieving state-of-the-art cloud and cloud shadow identification with OmniCloudMask

Deep learning models are widely used to extract features and insights from remotely sensed imagery. However, these models typically perform optimally when applied to the same sensor, resolution and imagery processing level as used during their training, and are rarely used or evaluated on out-of-dom...

Full description

Saved in:
Bibliographic Details
Published inRemote sensing of environment Vol. 322; p. 114694
Main Authors Wright, Nicholas, Duncan, John M.A., Callow, J. Nik, Thompson, Sally E., George, Richard J.
Format Journal Article
LanguageEnglish
Published Elsevier Inc 15.05.2025
Subjects
Online AccessGet full text
ISSN0034-4257
1879-0704
DOI10.1016/j.rse.2025.114694

Cover

More Information
Summary:Deep learning models are widely used to extract features and insights from remotely sensed imagery. However, these models typically perform optimally when applied to the same sensor, resolution and imagery processing level as used during their training, and are rarely used or evaluated on out-of-domain data. This limitation results in duplication of efforts in collecting similar training datasets from different satellites to train sensor-specific models. Here, we introduce a range of techniques to train deep learning models that generalise across various sensors, resolutions, and processing levels. We applied this approach to train OmniCloudMask (OCM), a sensor-agnostic deep learning model that segments clouds and cloud shadow. OCM demonstrates robust state-of-the-art performance across various satellite platforms when classifying clear, cloud, and shadow classes, with balanced overall accuracy values across: Landsat (91.5 % clear, 91.5 % cloud, and 75.2 % shadow); Sentinel-2 (92.2 % clear, 91.2 % cloud, and 80.5 % shadow); and PlanetScope (96.9 % clear, 98.8 % cloud, and 97.4 % shadow). OCM achieves this accuracy while only being trained on a single Sentinel-2 dataset, employing spectral normalisation and mixed resolution training to address the spectral and spatial differences between satellite platforms. This approach allows the model to effectively handle imagery from different sensors within the 10 m to 50 m resolution range, as well as higher resolution imagery that has been resampled to 10 m. The OCM library is available as an open source Python package on PyPI. [Display omitted] •Sensor-agnostic approach for training remote sensing deep learning models.•Spectral alignment with dynamic Z-score normalisation.•Range of spatial resolutions from 10 m to 50 m.•Open source OmniCloudMask model for cloud and cloud shadow detection.•State-of-the-art accuracy on datasets from Sentinel-2, Landsat 8 and PlanetScope.
ISSN:0034-4257
1879-0704
DOI:10.1016/j.rse.2025.114694