{"title":"Improved contrastive learning model via identification of false-negatives in self-supervised learning","authors":"Joonsun Auh, Changsik Cho, Seon-tae Kim","doi":"10.4218/etrij.2023-0285","DOIUrl":null,"url":null,"abstract":"<p>Self-supervised learning is a method that learns the data representation through unlabeled data. It is efficient because it learns from large-scale unlabeled data and through continuous research, performance comparable to supervised learning has been reached. Contrastive learning, a type of self-supervised learning algorithm, utilizes data similarity to perform instance-level learning within an embedding space. However, it suffers from the problem of false-negatives, which are the misclassification of data class during training the data representation. They result in loss of information and deteriorate the performance of the model. This study employed cosine similarity and temperature simultaneously to identify false-negatives and mitigate their impact to improve the performance of the contrastive learning model. The proposed method exhibited a performance improvement of up to 2.7% compared with the existing algorithm on the CIFAR-100 dataset. Improved performance on other datasets such as CIFAR-10 and ImageNet was also observed.</p>","PeriodicalId":11901,"journal":{"name":"ETRI Journal","volume":"46 6","pages":"1020-1029"},"PeriodicalIF":1.3000,"publicationDate":"2024-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.4218/etrij.2023-0285","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ETRI Journal","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.4218/etrij.2023-0285","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Self-supervised learning is a method that learns the data representation through unlabeled data. It is efficient because it learns from large-scale unlabeled data and through continuous research, performance comparable to supervised learning has been reached. Contrastive learning, a type of self-supervised learning algorithm, utilizes data similarity to perform instance-level learning within an embedding space. However, it suffers from the problem of false-negatives, which are the misclassification of data class during training the data representation. They result in loss of information and deteriorate the performance of the model. This study employed cosine similarity and temperature simultaneously to identify false-negatives and mitigate their impact to improve the performance of the contrastive learning model. The proposed method exhibited a performance improvement of up to 2.7% compared with the existing algorithm on the CIFAR-100 dataset. Improved performance on other datasets such as CIFAR-10 and ImageNet was also observed.
期刊介绍:
ETRI Journal is an international, peer-reviewed multidisciplinary journal published bimonthly in English. The main focus of the journal is to provide an open forum to exchange innovative ideas and technology in the fields of information, telecommunications, and electronics.
Key topics of interest include high-performance computing, big data analytics, cloud computing, multimedia technology, communication networks and services, wireless communications and mobile computing, material and component technology, as well as security.
With an international editorial committee and experts from around the world as reviewers, ETRI Journal publishes high-quality research papers on the latest and best developments from the global community.