PhilEO Bench: Evaluating Geo-Spatial Foundation Models (2401.04464v2)
Abstract: Massive amounts of unlabelled data are captured by Earth Observation (EO) satellites, with the Sentinel-2 constellation generating 1.6 TB of data daily. This makes Remote Sensing a data-rich domain well suited to Machine Learning (ML) solutions. However, a bottleneck in applying ML models to EO is the lack of annotated data as annotation is a labour-intensive and costly process. As a result, research in this domain has focused on Self-Supervised Learning and Foundation Model approaches. This paper addresses the need to evaluate different Foundation Models on a fair and uniform benchmark by introducing the PhilEO Bench, a novel evaluation framework for EO Foundation Models. The framework comprises of a testbed and a novel 400 GB Sentinel-2 dataset containing labels for three downstream tasks, building density estimation, road segmentation, and land cover classification. We present experiments using our framework evaluating different Foundation Models, including Prithvi and SatMAE, at multiple n-shots and convergence rates.
- Yi Wang et al., “Self-supervised learning in remote sensing: A review,” IEEE Geosc & Rem Sen Mag, 2022.
- European Space Agency (ESA), “Sentinel-2 Operations,” Enabling and Support, Operations, http://www.esa.int/Enabling_Support/Operations/Sentinel-2_operations.
- Oscar Manas et al., “Seasonal contrast: Unsupervised pre-training from uncurated remote sensing data,” in Proc. ICCV, 2021.
- “Geography-aware self-supervised learning,” arxiv:2011.09980, 2021.
- Yi Wang et al., “SSL4EO-S12: A large-scale multi-modal, multi-temporal dataset for self-supervised learning in Earth Observation,” arxiv:2211.07044, 2022.
- “An agenda for multimodal Foundation Models for Earth Observation,” In Proc. IGARSS, 2023.
- “Foundation Models for Generalist Geospatial Artificial Intelligence,” arxiv:2310.18660, 2023.
- “SatMAE: Pre-training Transformers for temporal and multi-spectral satellite imagery,” arxiv:2207.08051, 2023.
- A. Dosovitskiy et al., “Discriminative unsupervised feature learning with exemplar CNNs,” Trans. PAMI, 2016.
- “Multi-task self-supervised visual learning,” in Proc. ICCV, 2017.
- “Embedding Earth: Self-supervised contrastive pre-training for dense land cover classification,” arxiv:2203.06041, 2022.
- Yang et al. Yu, “An unsupervised convolutional feature fusion network for deep representation of remote sensing images,” IEEE Geoscience Rem Sens Letters, 2018.
- Paul Berg et al., “Self-supervised learning for scene classification in remote sensing: Current state of the art and perspectives,” Remote Sensing, 2022.
- D. Tuia et al., “Artificial Intelligence to advance Earth observation: A perspective,” arxiv:2305.08413, 2023.
- Favyen Bastani et al., “Satlas: A large-scale, multi-task dataset for RS image understanding,” In ICCV, 2023.
- Matias Mendieta et al., “Towards geospatial foundation models via continual pretraining,” In Proc. ICCV, 2023.
- Qian Shi et al., “Globe230k: A benchmark dense-pixel annotation dataset for global land cover mapping,” J Rem Sen AAAS, 2023.
- Gencer Sumbul et al., “BigEarthNet-MM: A large scale multi-modal multi-label benchmark archive for RS image classification,” IEEE Geosc Rem Sens Mag, 2021.
- Xenia Ivashkovych et al., “CORSA deep Earth Observation semantic compression applied to flood detection,” ESA OBPDC workshop, 2022.
- J. Irvin et al., “USat: A unified self-supervised encoder for multi-sensor satellite,” arXiv:2312.02199, 2023.
- Tung Nguyen et al., “ClimaX: A foundation model for weather and climate,” arxiv:2301.10343, 2023.
- Kaiming He et al., “Masked autoencoders are scalable vision learners,” In Proc. CVPR, 2022.
- “Scale-MAE: A scale-aware masked autoencoder for multiscale geospatial representation learning,” arxiv:2212.14532, 2022.
- G. Tseng et al., “Lightweight, pre-trained Transformers for remote sensing timeseries,” arxiv:2304.14065, 2023.
- Utkarsh Mall et al., “Change-aware sampling and contrastive learning for satellite images,” in Proc. CVPR, 2023.
- A. Lacoste et al., “Geo-Bench: Toward Foundation Models for Earth monitoring,” arxiv:2306.03831, 2023.
- Wenyuan Li et al., “Geographical knowledge-driven representation learning for remote sensing images,” IEEE Trans Geoscience and Rem Sens, 60:1-16, 2022.
- “The trade-off between universality and label efficiency of representations from contrastive learning,” in Proc. ICLR, 2023.
- K. Klemmer et al., “SatCLIP: Global, general-purpose location embeddings,” arXiv:2311.17179, 2023.
- Isaac Corley et al., “Revisiting pre-trained remote sensing model benchmarks: Resizing and normalization matters,” arXiv:2305.13456, 2023.
- Danfeng Hong et al., “SpectralGPT: Spectral foundation model,” arXiv:2311.07113, 2023.
- Randall Balestriero et al., “A cookbook of self-supervised learning,” arXiv:2304.12210, 2023.
- O. Ronneberger et al., “U-Net: Convolutional networks for biomedical image segmentation,” In MICCAI, 2015.
- Ilya O Tolstikhin et al., “MLP-Mixer: An all-MLP architecture for vision,” In Proc. NeurIPS, 2021.
- A. Dosovitskiy et al., “An image is worth 16x16 words: Transformers for recognition at scale,” In ICLR, 2021.
- M. Kottek et al., “World map of the Köppen-Geiger climate classification,” Meteor Z, v. 15, p. 259-263, 2006.
- Casper Fibaek (4 papers)
- Luke Camilleri (3 papers)
- Andreas Luyts (3 papers)
- Nikolaos Dionelis (16 papers)
- Bertrand Le Saux (59 papers)