计算机科学
人工智能
机器学习
利用
训练集
标记数据
监督学习
任务(项目管理)
变压器
领域(数学分析)
半监督学习
共同训练
模式识别(心理学)
数据挖掘
人工神经网络
数学分析
物理
计算机安全
管理
数学
量子力学
电压
经济
作者
Ivica Dimitrovski,Ivan Kitanovski,Nikola Simidjievski,Dragi Kocev
出处
期刊:IEEE Geoscience and Remote Sensing Letters
[Institute of Electrical and Electronics Engineers]
日期:2024-01-01
卷期号:21: 1-5
标识
DOI:10.1109/lgrs.2024.3352926
摘要
We investigate the utility of in-domain self-supervised pre-training of vision models in the analysis of remote sensing imagery. Self-supervised learning (SSL) has emerged as a promising approach for remote sensing image classification due to its ability to exploit large amounts of unlabeled data. Unlike traditional supervised learning, SSL aims to learn representations of data without the need for explicit labels. This is achieved by formulating auxiliary tasks that can be used for pre-training models before fine-tuning them on a given downstream task. A common approach in practice to SSL pre-training is utilizing standard pre-training datasets, such as ImageNet. While relevant, such a general approach can have a sub-optimal influence on the downstream performance of models, especially on tasks from challenging domains such as remote sensing. In this paper, we analyze the effectiveness of SSL pre-training by employing the iBOT framework coupled with Vision transformers trained on Million-AID, a large and unlabeled remote sensing dataset. We present a comprehensive study of different self-supervised pre-training strategies and evaluate their effect across 14 downstream datasets with diverse properties. Our results demonstrate that leveraging large in-domain datasets for self-supervised pre-training consistently leads to improved predictive downstream performance, compared to the standard approaches found in practice.
科研通智能强力驱动
Strongly Powered by AbleSci AI