Dynamically Scaled Temperature in Self-Supervised Contrastive Learning

Siladittya Manna;Soumitri Chattopadhyay;Rakesh Dey;Umapada Pal;Saumik Bhattacharya
{"title":"Dynamically Scaled Temperature in Self-Supervised Contrastive Learning","authors":"Siladittya Manna;Soumitri Chattopadhyay;Rakesh Dey;Umapada Pal;Saumik Bhattacharya","doi":"10.1109/TAI.2024.3524979","DOIUrl":null,"url":null,"abstract":"In contemporary self-supervised contrastive algorithms such as SimCLR and MoCo, the task of balancing attraction between two semantically similar samples and repulsion between two samples of different classes is primarily affected by the presence of hard negative samples. While the InfoNCE loss has been shown to impose penalties based on hardness, the temperature hyperparameter is the key to regulate the penalties and the tradeoff between uniformity and tolerance. In this work, we focus our attention on improving the performance of InfoNCE loss in self-supervised learning by proposing a novel cosine similarity dependent temperature scaling function to effectively optimize the distribution of the samples in the feature space. We also provide mathematical analyzes to support the construction of such a dynamically scaled temperature function. Experimental evidence shows that the proposed framework outperforms the contrastive loss-based SSL algorithms.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 6","pages":"1502-1512"},"PeriodicalIF":0.0000,"publicationDate":"2025-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10820841/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In contemporary self-supervised contrastive algorithms such as SimCLR and MoCo, the task of balancing attraction between two semantically similar samples and repulsion between two samples of different classes is primarily affected by the presence of hard negative samples. While the InfoNCE loss has been shown to impose penalties based on hardness, the temperature hyperparameter is the key to regulate the penalties and the tradeoff between uniformity and tolerance. In this work, we focus our attention on improving the performance of InfoNCE loss in self-supervised learning by proposing a novel cosine similarity dependent temperature scaling function to effectively optimize the distribution of the samples in the feature space. We also provide mathematical analyzes to support the construction of such a dynamically scaled temperature function. Experimental evidence shows that the proposed framework outperforms the contrastive loss-based SSL algorithms.
自监督对比学习中的动态缩放温度
在当代的自监督对比算法中,如SimCLR和MoCo,平衡两个语义相似样本之间的吸引力和两个不同类别样本之间的排斥力的任务主要受到硬负样本存在的影响。虽然已经证明InfoNCE损耗会基于硬度施加惩罚,但温度超参数是调节惩罚以及在均匀性和公差之间进行权衡的关键。在这项工作中,我们通过提出一种新的余弦相似度依赖的温度缩放函数来有效地优化样本在特征空间中的分布,从而提高自监督学习中InfoNCE损失的性能。我们还提供了数学分析来支持这种动态缩放温度函数的构建。实验结果表明,该框架优于基于对比损失的SSL算法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
7.70
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信