elib
DLR-Header
DLR-Logo -> http://www.dlr.de
DLR Portal Home | Imprint | Privacy Policy | Accessibility | Contact | Deutsch
Fontsize: [-] Text [+]

Multi-Label Guided Soft Contrastive Learning for Efficient Earth Observation Pretraining

Wang, Yi and Albrecht, Conrad M and Zhu, Xiao Xiang (2024) Multi-Label Guided Soft Contrastive Learning for Efficient Earth Observation Pretraining. IEEE Transactions on Geoscience and Remote Sensing, 62, p. 5644516. IEEE - Institute of Electrical and Electronics Engineers. doi: 10.1109/TGRS.2024.3466896. ISSN 0196-2892.

[img] PDF - Preprint version (submitted draft)
5MB

Official URL: https://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=10726860

Abstract

Self-supervised pretraining on large-scale satellite data has raised great interest in building Earth observation (EO) foundation models. However, many important resources beyond pure satellite imagery, such as land-cover-land-use products that provide free global semantic information, as well as vision foundation models that hold strong knowledge of the natural world, are not widely studied. In this work, we show these free additional resources not only help resolve common contrastive learning bottlenecks, but also significantly boost the efficiency and effectiveness of EO pretraining. Specifically, we first propose soft contrastive learning that optimizes cross-scene soft similarity based on land-cover-generated multi-label supervision, naturally solving the issue of multiple positive samples and too strict positive matching in complex scenes. Second, we revisit and explore cross-domain continual pretraining for both multispectral and SAR imagery, building efficient EO foundation models from strongest vision models such as DINOv2. Adapting simple weight-initialization and Siamese masking strategies into our soft contrastive learning framework, we demonstrate impressive continual pretraining performance even when the input modalities are not aligned. Without prohibitive training, we produce multispectral and SAR foundation models that achieve significantly better results in 10 out of 11 downstream tasks than most existing SOTA models. For example, our ResNet50/ViT-S achieve 84.8/85.0 linear probing mAP scores on BigEarthNet-10% which are better than most existing ViT-L models; under the same setting, our ViT-B sets a new record of 86.8 in multispectral, and 82.5 in SAR, the latter even better than many multispectral models

Item URL in elib:https://elib.dlr.de/207106/
Document Type:Article
Title:Multi-Label Guided Soft Contrastive Learning for Efficient Earth Observation Pretraining
Authors:
AuthorsInstitution or Email of AuthorsAuthor's ORCID iDORCID Put Code
Wang, YiUNSPECIFIEDUNSPECIFIEDUNSPECIFIED
Albrecht, Conrad MUNSPECIFIEDhttps://orcid.org/0009-0009-2422-7289UNSPECIFIED
Zhu, Xiao XiangUNSPECIFIEDUNSPECIFIEDUNSPECIFIED
Date:2024
Journal or Publication Title:IEEE Transactions on Geoscience and Remote Sensing
Refereed publication:Yes
Open Access:Yes
Gold Open Access:No
In SCOPUS:Yes
In ISI Web of Science:Yes
Volume:62
DOI:10.1109/TGRS.2024.3466896
Page Range:p. 5644516
Publisher:IEEE - Institute of Electrical and Electronics Engineers
ISSN:0196-2892
Status:Published
Keywords:weakly supervised learning, contrastive self-supervised learning, multispectral, SAR, geospatial foundation model
HGF - Research field:Aeronautics, Space and Transport
HGF - Program:Space
HGF - Program Themes:Earth Observation
DLR - Research area:Raumfahrt
DLR - Program:R EO - Earth Observation
DLR - Research theme (Project):R - SAR methods, R - Optical remote sensing, R - Artificial Intelligence
Location: Oberpfaffenhofen
Institutes and Institutions:Remote Sensing Technology Institute > EO Data Science
Deposited By: Albrecht, Conrad M
Deposited On:07 Oct 2024 10:19
Last Modified:09 Dec 2025 13:09

Repository Staff Only: item control page

Browse
Search
Help & Contact
Information
OpenAIRE Validator logo electronic library is running on EPrints 3.3.12
Website and database design: Copyright © German Aerospace Center (DLR). All rights reserved.