Ning Li , Junjie Hou , Wenjiao Zhang , Yanan Zhuang , Qianqian Xu , Haohan Yong
{"title":"A speech emotion recognition method based on DST-GDCC and text-to-speech data augmentation","authors":"Ning Li , Junjie Hou , Wenjiao Zhang , Yanan Zhuang , Qianqian Xu , Haohan Yong","doi":"10.1016/j.dsp.2025.105636","DOIUrl":null,"url":null,"abstract":"<div><div>Speech Emotion Recognition (SER) is a critical component of human-machine interaction, yet it confronts two fundamental challenges: limited feature extraction capabilities and data scarcity. This paper proposes a unified framework that synergistically addresses both issues through the co-design of a novel SER model and a high-quality data augmentation strategy. At its core, the Deformable Speech Transformer (DST) and the Gated Dilation Causal Convolution (GDCC) are introduced, which are combined to form the DST-GDCC model for superior feature extraction. The DST component adaptively captures multi-granular acoustic features, while the GDCC module explicitly models the spatiotemporal causality of speech emotions. However, the full potential of such an advanced model is often constrained by scarce training data. To overcome this limitation, a Text-to-Speech (TTS) data augmentation method is incorporated, leveraging a pre-trained GPT-SoVITS model to synthesize high-fidelity, emotion-conditioned speech samples. Crucially, these two components form a virtuous cycle: the powerful discriminative ability of the DST-GDCC model is leveraged in a dual-stage screening mechanism to ensure the quality of the synthetic data, while the expanded, high-quality dataset, in turn, enables the model to realize its full potential. Experimental results demonstrate the framework's effectiveness. The DST-GDCC model itself achieves significant accuracy improvements over baselines (2.66% on IEMOCAP, 5.02% on MELD, 5.83% on CASIA). More importantly, the synergistic integration with TTS data augmentation yields further gains of 3.13% on IEMOCAP and 3.33% on CASIA, validating the framework's capability to systematically elevate SER performance.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"168 ","pages":"Article 105636"},"PeriodicalIF":3.0000,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Digital Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S105120042500658X","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Speech Emotion Recognition (SER) is a critical component of human-machine interaction, yet it confronts two fundamental challenges: limited feature extraction capabilities and data scarcity. This paper proposes a unified framework that synergistically addresses both issues through the co-design of a novel SER model and a high-quality data augmentation strategy. At its core, the Deformable Speech Transformer (DST) and the Gated Dilation Causal Convolution (GDCC) are introduced, which are combined to form the DST-GDCC model for superior feature extraction. The DST component adaptively captures multi-granular acoustic features, while the GDCC module explicitly models the spatiotemporal causality of speech emotions. However, the full potential of such an advanced model is often constrained by scarce training data. To overcome this limitation, a Text-to-Speech (TTS) data augmentation method is incorporated, leveraging a pre-trained GPT-SoVITS model to synthesize high-fidelity, emotion-conditioned speech samples. Crucially, these two components form a virtuous cycle: the powerful discriminative ability of the DST-GDCC model is leveraged in a dual-stage screening mechanism to ensure the quality of the synthetic data, while the expanded, high-quality dataset, in turn, enables the model to realize its full potential. Experimental results demonstrate the framework's effectiveness. The DST-GDCC model itself achieves significant accuracy improvements over baselines (2.66% on IEMOCAP, 5.02% on MELD, 5.83% on CASIA). More importantly, the synergistic integration with TTS data augmentation yields further gains of 3.13% on IEMOCAP and 3.33% on CASIA, validating the framework's capability to systematically elevate SER performance.
期刊介绍:
Digital Signal Processing: A Review Journal is one of the oldest and most established journals in the field of signal processing yet it aims to be the most innovative. The Journal invites top quality research articles at the frontiers of research in all aspects of signal processing. Our objective is to provide a platform for the publication of ground-breaking research in signal processing with both academic and industrial appeal.
The journal has a special emphasis on statistical signal processing methodology such as Bayesian signal processing, and encourages articles on emerging applications of signal processing such as:
• big data• machine learning• internet of things• information security• systems biology and computational biology,• financial time series analysis,• autonomous vehicles,• quantum computing,• neuromorphic engineering,• human-computer interaction and intelligent user interfaces,• environmental signal processing,• geophysical signal processing including seismic signal processing,• chemioinformatics and bioinformatics,• audio, visual and performance arts,• disaster management and prevention,• renewable energy,