{"title":"Reproducible and generalizable speech emotion recognition via an Intelligent Fusion Network","authors":"Huiyun Zhang , Puyang Zhao , Gaigai Tang , Zongjin Li , Zhu Yuan","doi":"10.1016/j.bspc.2025.107996","DOIUrl":null,"url":null,"abstract":"<div><div>Speech emotion recognition (SER) is a critical aspect of enhancing the naturalness and effectiveness of Human-computer interaction systems. Despite substantial progress through deep learning techniques, challenges remain, particularly concerning model performance, reproducibility, and generalization. To address these challenges, we propose the Intelligent Fusion Network (IFN), a novel framework designed to improve emotion recognition by leveraging an isomorphic architecture and attention-based mechanisms. The IFN framework consists of five key components: an input processing layer, a feature mapping module, a dual attention mechanism, a convolutional feature refinement module, and a multiplicative fusion module, culminating in an output layer. In addition, we introduce a robust methodology for quantifying and assessing the reproducibility of deep learning models, ensuring consistent and reliable evaluations. Extensive experiments conducted across six benchmark datasets—EMODB, CASIA, SAVEE, BodEMODB, IEMOCAP, and ESD—demonstrate the superior performance of IFN. Specifically, IFN achieves an accuracy of 96.31 % on the ESD dataset, surpassing the leading baseline by 2.70 %. On the more challenging IEMOCAP dataset, IFN attains an accuracy of 64.32 %, highlighting its ability to generalize effectively across diverse datasets. Furthermore, IFN demonstrates exceptional reproducibility, with general and correct reproducibility rates of 86.69 % and 86.34 %, respectively, at k = 10 on the ESD dataset, significantly outperforming existing approaches. These results highlight IFN as a highly reliable and effective solution for advancing SER, offering the potential to enable more intuitive and efficient Human-Computer Interaction (HCI) systems.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"109 ","pages":"Article 107996"},"PeriodicalIF":4.9000,"publicationDate":"2025-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomedical Signal Processing and Control","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1746809425005075","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Speech emotion recognition (SER) is a critical aspect of enhancing the naturalness and effectiveness of Human-computer interaction systems. Despite substantial progress through deep learning techniques, challenges remain, particularly concerning model performance, reproducibility, and generalization. To address these challenges, we propose the Intelligent Fusion Network (IFN), a novel framework designed to improve emotion recognition by leveraging an isomorphic architecture and attention-based mechanisms. The IFN framework consists of five key components: an input processing layer, a feature mapping module, a dual attention mechanism, a convolutional feature refinement module, and a multiplicative fusion module, culminating in an output layer. In addition, we introduce a robust methodology for quantifying and assessing the reproducibility of deep learning models, ensuring consistent and reliable evaluations. Extensive experiments conducted across six benchmark datasets—EMODB, CASIA, SAVEE, BodEMODB, IEMOCAP, and ESD—demonstrate the superior performance of IFN. Specifically, IFN achieves an accuracy of 96.31 % on the ESD dataset, surpassing the leading baseline by 2.70 %. On the more challenging IEMOCAP dataset, IFN attains an accuracy of 64.32 %, highlighting its ability to generalize effectively across diverse datasets. Furthermore, IFN demonstrates exceptional reproducibility, with general and correct reproducibility rates of 86.69 % and 86.34 %, respectively, at k = 10 on the ESD dataset, significantly outperforming existing approaches. These results highlight IFN as a highly reliable and effective solution for advancing SER, offering the potential to enable more intuitive and efficient Human-Computer Interaction (HCI) systems.
期刊介绍:
Biomedical Signal Processing and Control aims to provide a cross-disciplinary international forum for the interchange of information on research in the measurement and analysis of signals and images in clinical medicine and the biological sciences. Emphasis is placed on contributions dealing with the practical, applications-led research on the use of methods and devices in clinical diagnosis, patient monitoring and management.
Biomedical Signal Processing and Control reflects the main areas in which these methods are being used and developed at the interface of both engineering and clinical science. The scope of the journal is defined to include relevant review papers, technical notes, short communications and letters. Tutorial papers and special issues will also be published.