MIND YOUR AUGMENTATION: THE KEY TO DECOUPLING DENSE SELF-SUPERVISED LEARNING

Congpei Qiu, Tong Zhang, Yanhao Wu, Wei Ke, Mathieu Salzmann, Sabine Süsstrunk

研究成果: Paper同行評審

摘要

Dense Self-Supervised Learning (SSL) creates positive pairs by building positive paired regions or points, thereby aiming to preserve local features, for example of individual objects. However, existing approaches tend to couple objects by leaking information from the neighboring contextual regions when the pairs have a limited overlap. In this paper, we first quantitatively identify and confirm the existence of such a coupling phenomenon. We then address it by developing a remarkably simple yet highly effective solution comprising a novel augmentation method, Region Collaborative Cutout (RCC), and a corresponding decoupling branch. Importantly, our design is versatile and can be seamlessly integrated into existing SSL frameworks, whether based on Convolutional Neural Networks (CNNs) or Vision Transformers (ViTs). We conduct extensive experiments, incorporating our solution into two CNN-based and two ViT-based methods, with results confirming the effectiveness of our approach. Moreover, we provide empirical evidence that our method significantly contributes to the disentanglement of feature representations among objects, both in quantitative and qualitative terms.

原文English
出版狀態Published - 2024
對外發佈
事件12th International Conference on Learning Representations, ICLR 2024 - Hybrid, Vienna, Austria
持續時間: 7 5月 202411 5月 2024

Conference

Conference12th International Conference on Learning Representations, ICLR 2024
國家/地區Austria
城市Hybrid, Vienna
期間7/05/2411/05/24

指紋

深入研究「MIND YOUR AUGMENTATION: THE KEY TO DECOUPLING DENSE SELF-SUPERVISED LEARNING」主題。共同形成了獨特的指紋。

引用此