Self-Supervised Visual Representation Learning via Residual Momentum
Self-supervised learning (SSL) has emerged as a promising approach for learning representations from unlabeled data. Momentum-based contrastive frameworks such as MoCo-v3 have shown remarkable success among the many SSL methods proposed in recent years. However, a significant gap in encoder represen...
Main Authors: | , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2023-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10287941/ |
_version_ | 1797648645681577984 |
---|---|
author | Trung Xuan Pham Axi Niu Kang Zhang Tee Joshua Tian Jin Ji Woo Hong Chang D. Yoo |
author_facet | Trung Xuan Pham Axi Niu Kang Zhang Tee Joshua Tian Jin Ji Woo Hong Chang D. Yoo |
author_sort | Trung Xuan Pham |
collection | DOAJ |
description | Self-supervised learning (SSL) has emerged as a promising approach for learning representations from unlabeled data. Momentum-based contrastive frameworks such as MoCo-v3 have shown remarkable success among the many SSL methods proposed in recent years. However, a significant gap in encoder representation exists between the online encoder (student) and the momentum encoder (teacher) in these frameworks, limiting the performance on downstream tasks. We identify this gap as a bottleneck often overlooked in existing frameworks and propose “residual momentum” that explicitly reduces the gap during training to encourage the student to learn representations closer to the teacher’s. We also reveal that a similar technique, knowledge distillation (KD), to reduce the distribution gap with cross-entropy-based loss in supervised learning is useless in the SSL context and demonstrate that the intra-representation gap measured by cosine similarity is crucial for EMA-based SSLs. Extensive experiments on different benchmark datasets and architectures demonstrate the superiority of our method compared to state-of-the-art contrastive learning baselines. Specifically, our method outperforms MoCo-v3 0.7% top-1 in ImageNet, 2.82% on CIFAR-100, 1.8% AP, and 3.0% AP75 on VOC detection pre-trained on the COCO dataset; it also improves DenseCL with 0.5% AP (800ep) and 0.6% AP75 (1600ep). Our work highlights the importance of reducing the teacher-student intra-gap in momentum-based contrastive learning frameworks and provides a practical solution for improving the quality of learned representations. |
first_indexed | 2024-03-11T15:33:57Z |
format | Article |
id | doaj.art-40a6e2d1b58f4821815151218f40b9d6 |
institution | Directory Open Access Journal |
issn | 2169-3536 |
language | English |
last_indexed | 2024-03-11T15:33:57Z |
publishDate | 2023-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj.art-40a6e2d1b58f4821815151218f40b9d62023-10-26T23:01:26ZengIEEEIEEE Access2169-35362023-01-011111670611672010.1109/ACCESS.2023.332584210287941Self-Supervised Visual Representation Learning via Residual MomentumTrung Xuan Pham0https://orcid.org/0000-0003-4177-7054Axi Niu1https://orcid.org/0000-0001-5238-9917Kang Zhang2https://orcid.org/0000-0003-2761-9383Tee Joshua Tian Jin3https://orcid.org/0009-0001-5119-2802Ji Woo Hong4https://orcid.org/0000-0002-3758-0307Chang D. Yoo5https://orcid.org/0000-0002-0756-7179School of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South KoreaSchool of Computer Science, Northwestern Polytechnical University, Xi’an, ChinaSchool of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South KoreaSchool of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South KoreaSchool of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South KoreaSchool of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon, South KoreaSelf-supervised learning (SSL) has emerged as a promising approach for learning representations from unlabeled data. Momentum-based contrastive frameworks such as MoCo-v3 have shown remarkable success among the many SSL methods proposed in recent years. However, a significant gap in encoder representation exists between the online encoder (student) and the momentum encoder (teacher) in these frameworks, limiting the performance on downstream tasks. We identify this gap as a bottleneck often overlooked in existing frameworks and propose “residual momentum” that explicitly reduces the gap during training to encourage the student to learn representations closer to the teacher’s. We also reveal that a similar technique, knowledge distillation (KD), to reduce the distribution gap with cross-entropy-based loss in supervised learning is useless in the SSL context and demonstrate that the intra-representation gap measured by cosine similarity is crucial for EMA-based SSLs. Extensive experiments on different benchmark datasets and architectures demonstrate the superiority of our method compared to state-of-the-art contrastive learning baselines. Specifically, our method outperforms MoCo-v3 0.7% top-1 in ImageNet, 2.82% on CIFAR-100, 1.8% AP, and 3.0% AP75 on VOC detection pre-trained on the COCO dataset; it also improves DenseCL with 0.5% AP (800ep) and 0.6% AP75 (1600ep). Our work highlights the importance of reducing the teacher-student intra-gap in momentum-based contrastive learning frameworks and provides a practical solution for improving the quality of learned representations.https://ieeexplore.ieee.org/document/10287941/Contrastive learningresidual momentumrepresentation learningself-supervised learningknowledge distillationteacher-student gap |
spellingShingle | Trung Xuan Pham Axi Niu Kang Zhang Tee Joshua Tian Jin Ji Woo Hong Chang D. Yoo Self-Supervised Visual Representation Learning via Residual Momentum IEEE Access Contrastive learning residual momentum representation learning self-supervised learning knowledge distillation teacher-student gap |
title | Self-Supervised Visual Representation Learning via Residual Momentum |
title_full | Self-Supervised Visual Representation Learning via Residual Momentum |
title_fullStr | Self-Supervised Visual Representation Learning via Residual Momentum |
title_full_unstemmed | Self-Supervised Visual Representation Learning via Residual Momentum |
title_short | Self-Supervised Visual Representation Learning via Residual Momentum |
title_sort | self supervised visual representation learning via residual momentum |
topic | Contrastive learning residual momentum representation learning self-supervised learning knowledge distillation teacher-student gap |
url | https://ieeexplore.ieee.org/document/10287941/ |
work_keys_str_mv | AT trungxuanpham selfsupervisedvisualrepresentationlearningviaresidualmomentum AT axiniu selfsupervisedvisualrepresentationlearningviaresidualmomentum AT kangzhang selfsupervisedvisualrepresentationlearningviaresidualmomentum AT teejoshuatianjin selfsupervisedvisualrepresentationlearningviaresidualmomentum AT jiwoohong selfsupervisedvisualrepresentationlearningviaresidualmomentum AT changdyoo selfsupervisedvisualrepresentationlearningviaresidualmomentum |