基于自适应超球上近邻分布的神经网络分类器学习

Xiaojing Zhang;Shuangrong Liu;Lin Wang;Bo Yang;Jiawei Fan
{"title":"基于自适应超球上近邻分布的神经网络分类器学习","authors":"Xiaojing Zhang;Shuangrong Liu;Lin Wang;Bo Yang;Jiawei Fan","doi":"10.1109/TAI.2024.3477436","DOIUrl":null,"url":null,"abstract":"In this study, the adaptive hypersphere nearest neighbors (ASNN) method is proposed as an optimization framework to enhance the generalization performance of neural network classifiers. In terms of the classification task, the neural network draws decision boundaries by constructing the discriminative features of samples. To learn those features, attributed to the flexibility and separability, the pair-wise constraint-based methods that consist of the pair-wise loss and an embedding space (e.g., hypersphere space) have gained considerable attention over the last decade. Despite their success, pair-wise constraint-based methods still suffer from premature convergence or divergence problems, driven by two main challenges. 1) The poor scalability of the embedding space constrains the variety of the distribution of embedded samples, thereby increasing the optimization difficulty. 2) It is hard to select suitable positive/negative pairs during the training. In order to address the aforementioned problems, we propose an adaptive hypersphere nearest neighbors method. On the one hand, we improve the scalability of features via a scale-adaptive hypersphere embedding space. On the other hand, we introduce a neighborhood-based probability loss, which magnifies the difference between pairs and enhances the discriminative power of features generated by the neural networks based on the nearest neighbor-based pairing strategy. Experiments on UCI datasets and image recognition tasks demonstrate that the proposed ASNN not only achieves improved intraclass consistency and interclass separability of samples, but also outperforms its competitive counterparts.","PeriodicalId":73305,"journal":{"name":"IEEE transactions on artificial intelligence","volume":"6 1","pages":"234-249"},"PeriodicalIF":0.0000,"publicationDate":"2024-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Learning Neural Network Classifiers by Distributing Nearest Neighbors on Adaptive Hypersphere\",\"authors\":\"Xiaojing Zhang;Shuangrong Liu;Lin Wang;Bo Yang;Jiawei Fan\",\"doi\":\"10.1109/TAI.2024.3477436\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this study, the adaptive hypersphere nearest neighbors (ASNN) method is proposed as an optimization framework to enhance the generalization performance of neural network classifiers. In terms of the classification task, the neural network draws decision boundaries by constructing the discriminative features of samples. To learn those features, attributed to the flexibility and separability, the pair-wise constraint-based methods that consist of the pair-wise loss and an embedding space (e.g., hypersphere space) have gained considerable attention over the last decade. Despite their success, pair-wise constraint-based methods still suffer from premature convergence or divergence problems, driven by two main challenges. 1) The poor scalability of the embedding space constrains the variety of the distribution of embedded samples, thereby increasing the optimization difficulty. 2) It is hard to select suitable positive/negative pairs during the training. In order to address the aforementioned problems, we propose an adaptive hypersphere nearest neighbors method. On the one hand, we improve the scalability of features via a scale-adaptive hypersphere embedding space. On the other hand, we introduce a neighborhood-based probability loss, which magnifies the difference between pairs and enhances the discriminative power of features generated by the neural networks based on the nearest neighbor-based pairing strategy. Experiments on UCI datasets and image recognition tasks demonstrate that the proposed ASNN not only achieves improved intraclass consistency and interclass separability of samples, but also outperforms its competitive counterparts.\",\"PeriodicalId\":73305,\"journal\":{\"name\":\"IEEE transactions on artificial intelligence\",\"volume\":\"6 1\",\"pages\":\"234-249\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-10-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE transactions on artificial intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10713256/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on artificial intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10713256/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

为了提高神经网络分类器的泛化性能,本文提出了自适应超球最近邻(ASNN)方法作为优化框架。在分类任务方面,神经网络通过构造样本的判别特征来绘制决策边界。为了学习这些特征,由于其灵活性和可分离性,由成对损失和嵌入空间(如超球空间)组成的基于成对约束的方法在过去十年中获得了相当大的关注。尽管它们取得了成功,但基于成对约束的方法仍然存在过早收敛或分歧的问题,这主要受到两个主要挑战的驱动。1)嵌入空间的可扩展性差,限制了嵌入样本分布的多样性,从而增加了优化难度。2)在训练过程中很难选择合适的正/负对。为了解决上述问题,我们提出了一种自适应超球最近邻方法。一方面,我们通过自适应尺度的超球嵌入空间提高了特征的可扩展性。另一方面,引入基于邻域的概率损失,放大了基于最近邻配对策略的神经网络对特征的差异,增强了神经网络对特征的判别能力。在UCI数据集和图像识别任务上的实验表明,该方法不仅提高了样本的类内一致性和类间可分离性,而且优于同类方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Learning Neural Network Classifiers by Distributing Nearest Neighbors on Adaptive Hypersphere
In this study, the adaptive hypersphere nearest neighbors (ASNN) method is proposed as an optimization framework to enhance the generalization performance of neural network classifiers. In terms of the classification task, the neural network draws decision boundaries by constructing the discriminative features of samples. To learn those features, attributed to the flexibility and separability, the pair-wise constraint-based methods that consist of the pair-wise loss and an embedding space (e.g., hypersphere space) have gained considerable attention over the last decade. Despite their success, pair-wise constraint-based methods still suffer from premature convergence or divergence problems, driven by two main challenges. 1) The poor scalability of the embedding space constrains the variety of the distribution of embedded samples, thereby increasing the optimization difficulty. 2) It is hard to select suitable positive/negative pairs during the training. In order to address the aforementioned problems, we propose an adaptive hypersphere nearest neighbors method. On the one hand, we improve the scalability of features via a scale-adaptive hypersphere embedding space. On the other hand, we introduce a neighborhood-based probability loss, which magnifies the difference between pairs and enhances the discriminative power of features generated by the neural networks based on the nearest neighbor-based pairing strategy. Experiments on UCI datasets and image recognition tasks demonstrate that the proposed ASNN not only achieves improved intraclass consistency and interclass separability of samples, but also outperforms its competitive counterparts.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
7.70
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信