Shuai Yuan, Lei Yu, Tian Yao, Tianya Mao, Wen Xie, Jiajie Wang
{"title":"SIMNet: an infrared image action recognition network based on similarity evaluation","authors":"Shuai Yuan, Lei Yu, Tian Yao, Tianya Mao, Wen Xie, Jiajie Wang","doi":"10.1007/s10043-025-00967-y","DOIUrl":null,"url":null,"abstract":"<p>Infrared sensors are widely used in human action recognition because of their low light influence and excellent privacy protection. However, the traditional deep learning networks and training or testing methods tend to fall into the trap of local optimum because of the similarity between infrared image classes and the lack of discriminative features such as texture and depth, and thus obtain poor recognition results. To address this issue, we propose a novel human action recognition method based on similarity evaluation. This method innovatively transforms the traditional training and testing (verification) mode. First, we use a feature-to-feature training method to make the network pay more attention to the behavioral information that distinguishes the classes. Second, we design a Integrate Channel Attention Module(ICA) to enable Siamese network to focus on the areas of interest. Finally, we propose the Multimodal Similarity Evaluation Module (MSE). The module aims to address the fuzzy matching problem of feature areas. The contrast experiment results show that our method outperforms existing mainstream methods on several benchmark datasets. The excellent accuracy provides an innovative method for addressing various problems related to high similarity between classes.</p>","PeriodicalId":722,"journal":{"name":"Optical Review","volume":"34 1","pages":""},"PeriodicalIF":1.1000,"publicationDate":"2025-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Optical Review","FirstCategoryId":"101","ListUrlMain":"https://doi.org/10.1007/s10043-025-00967-y","RegionNum":4,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"OPTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Infrared sensors are widely used in human action recognition because of their low light influence and excellent privacy protection. However, the traditional deep learning networks and training or testing methods tend to fall into the trap of local optimum because of the similarity between infrared image classes and the lack of discriminative features such as texture and depth, and thus obtain poor recognition results. To address this issue, we propose a novel human action recognition method based on similarity evaluation. This method innovatively transforms the traditional training and testing (verification) mode. First, we use a feature-to-feature training method to make the network pay more attention to the behavioral information that distinguishes the classes. Second, we design a Integrate Channel Attention Module(ICA) to enable Siamese network to focus on the areas of interest. Finally, we propose the Multimodal Similarity Evaluation Module (MSE). The module aims to address the fuzzy matching problem of feature areas. The contrast experiment results show that our method outperforms existing mainstream methods on several benchmark datasets. The excellent accuracy provides an innovative method for addressing various problems related to high similarity between classes.
期刊介绍:
Optical Review is an international journal published by the Optical Society of Japan. The scope of the journal is:
General and physical optics;
Quantum optics and spectroscopy;
Information optics;
Photonics and optoelectronics;
Biomedical photonics and biological optics;
Lasers;
Nonlinear optics;
Optical systems and technologies;
Optical materials and manufacturing technologies;
Vision;
Infrared and short wavelength optics;
Cross-disciplinary areas such as environmental, energy, food, agriculture and space technologies;
Other optical methods and applications.