{"title":"Large-margin Softmax loss using synthetic virtual class","authors":"Jiuzhou Chen , Xiangyang Huang , Shudong Zhang","doi":"10.1016/j.neunet.2025.108068","DOIUrl":null,"url":null,"abstract":"<div><div>The primary challenge of large-margin learning lies in designing classifiers with strong discriminative power. Although existing large margin methods have achieved success in various classification tasks, they often suffer from weak task generalization and imbalanced handling of easy and hard samples. In this paper, we propose a margin adaptive synthetic virtual Softmax loss (SV-Softmax), which dynamically generates virtual prototypes by synthesizing embedded features and their corresponding prototypes. These virtual prototypes can adaptively adjust the margin based on the spatial distribution of embedded features, promoting the proximity of embedded features to their corresponding prototypes and creating clear and discriminative decision boundaries. Furthermore, we introduce a virtual prototype insertion strategy based on hard sample mining, where different synthesis strategies are applied to correctly and incorrectly classified samples, emphasizing the importance of hard samples. SV-Softmax is plug-and-play with minimal computational complexity, without requiring feature or weight normalization nor relying on task-specific hyperparameter tuning. Extensive comparative experiments on multiple visual classification and face recognition datasets demonstrate that SV-Softmax achieves competitive or superior performance compared to nine state-of-the-art methods. The code available at: <span><span>https://github.com/10zhou/SV-Softmax</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"193 ","pages":"Article 108068"},"PeriodicalIF":6.3000,"publicationDate":"2025-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0893608025009487","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The primary challenge of large-margin learning lies in designing classifiers with strong discriminative power. Although existing large margin methods have achieved success in various classification tasks, they often suffer from weak task generalization and imbalanced handling of easy and hard samples. In this paper, we propose a margin adaptive synthetic virtual Softmax loss (SV-Softmax), which dynamically generates virtual prototypes by synthesizing embedded features and their corresponding prototypes. These virtual prototypes can adaptively adjust the margin based on the spatial distribution of embedded features, promoting the proximity of embedded features to their corresponding prototypes and creating clear and discriminative decision boundaries. Furthermore, we introduce a virtual prototype insertion strategy based on hard sample mining, where different synthesis strategies are applied to correctly and incorrectly classified samples, emphasizing the importance of hard samples. SV-Softmax is plug-and-play with minimal computational complexity, without requiring feature or weight normalization nor relying on task-specific hyperparameter tuning. Extensive comparative experiments on multiple visual classification and face recognition datasets demonstrate that SV-Softmax achieves competitive or superior performance compared to nine state-of-the-art methods. The code available at: https://github.com/10zhou/SV-Softmax.
期刊介绍:
Neural Networks is a platform that aims to foster an international community of scholars and practitioners interested in neural networks, deep learning, and other approaches to artificial intelligence and machine learning. Our journal invites submissions covering various aspects of neural networks research, from computational neuroscience and cognitive modeling to mathematical analyses and engineering applications. By providing a forum for interdisciplinary discussions between biology and technology, we aim to encourage the development of biologically-inspired artificial intelligence.