Hierarchical Transformer With Auxiliary Learning for Subject-Independent Respiration Emotion Recognition

IF 4.3 2区 综合性期刊 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC
Yong Wang;Chendong Xu;Weirui Na;Dongyu Liu;Jiuqi Yan;Shuai Yao;Qisong Wu
{"title":"Hierarchical Transformer With Auxiliary Learning for Subject-Independent Respiration Emotion Recognition","authors":"Yong Wang;Chendong Xu;Weirui Na;Dongyu Liu;Jiuqi Yan;Shuai Yao;Qisong Wu","doi":"10.1109/JSEN.2025.3587271","DOIUrl":null,"url":null,"abstract":"Respiration is modulated by human emotional activity. Emotion recognition using physiological signals has recently gained considerable attention. However, most existing studies primarily focus on using electroencephalogram (EEG) signals for emotion recognition. This article explores the potential of utilizing respiration signals collected by wearable devices for emotion recognition. We propose a hierarchical transformer model to effectively extract emotional information from respiration signals. Furthermore, we introduce gender classification as an auxiliary task to further improve the accuracy of emotion recognition. Specifically, a frame transformer is employed to capture emotional information across frames of respiration signals. The extracted frame-level features are subsequently fused with segment-level embeddings through a specially designed fusion layer. Next, separate segment transformers are employed for emotion and gender to extract segment-level information, with a cosine similarity loss applied to promote shared feature learning. Finally, distinct classifiers are used for emotion and gender classification. To evaluate the effectiveness of the proposed method, extensive experiments are conducted on the DEAP and MAHNOB-HCI datasets under a subject-independent setting. For the DEAP dataset, the average classification accuracies are 72.42% for valence and 73.91% for arousal, while for the MAHNOB-HCI dataset, they are 80.45% and 79.69%, respectively. Compared to other physiological signals, such as EEG, respiration signals exhibit comparable potential for emotion recognition.","PeriodicalId":447,"journal":{"name":"IEEE Sensors Journal","volume":"25 16","pages":"31290-31301"},"PeriodicalIF":4.3000,"publicationDate":"2025-07-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Sensors Journal","FirstCategoryId":"103","ListUrlMain":"https://ieeexplore.ieee.org/document/11080225/","RegionNum":2,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

Respiration is modulated by human emotional activity. Emotion recognition using physiological signals has recently gained considerable attention. However, most existing studies primarily focus on using electroencephalogram (EEG) signals for emotion recognition. This article explores the potential of utilizing respiration signals collected by wearable devices for emotion recognition. We propose a hierarchical transformer model to effectively extract emotional information from respiration signals. Furthermore, we introduce gender classification as an auxiliary task to further improve the accuracy of emotion recognition. Specifically, a frame transformer is employed to capture emotional information across frames of respiration signals. The extracted frame-level features are subsequently fused with segment-level embeddings through a specially designed fusion layer. Next, separate segment transformers are employed for emotion and gender to extract segment-level information, with a cosine similarity loss applied to promote shared feature learning. Finally, distinct classifiers are used for emotion and gender classification. To evaluate the effectiveness of the proposed method, extensive experiments are conducted on the DEAP and MAHNOB-HCI datasets under a subject-independent setting. For the DEAP dataset, the average classification accuracies are 72.42% for valence and 73.91% for arousal, while for the MAHNOB-HCI dataset, they are 80.45% and 79.69%, respectively. Compared to other physiological signals, such as EEG, respiration signals exhibit comparable potential for emotion recognition.
基于辅助学习的分层变换呼吸情绪识别
呼吸受人类情绪活动的调节。最近,利用生理信号进行情绪识别得到了相当大的关注。然而,大多数现有研究主要集中在利用脑电图信号进行情绪识别。本文探讨了利用可穿戴设备收集的呼吸信号进行情绪识别的潜力。我们提出了一种层次转换器模型来有效地从呼吸信号中提取情绪信息。此外,我们引入性别分类作为辅助任务,进一步提高情感识别的准确性。具体来说,采用帧转换器来捕获呼吸信号帧间的情感信息。提取的帧级特征随后通过特殊设计的融合层与段级嵌入融合。接下来,对情感和性别使用单独的片段变换来提取片段级信息,并使用余弦相似度损失来促进共享特征学习。最后,使用不同的分类器进行情感和性别分类。为了评估该方法的有效性,我们在独立于主题的DEAP和MAHNOB-HCI数据集上进行了大量的实验。对于DEAP数据,效价和觉醒的平均分类准确率分别为72.42%和73.91%,而对于MAHNOB-HCI数据,它们的平均分类准确率分别为80.45%和79.69%。与其他生理信号(如脑电图)相比,呼吸信号在情绪识别方面表现出相当的潜力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
IEEE Sensors Journal
IEEE Sensors Journal 工程技术-工程:电子与电气
CiteScore
7.70
自引率
14.00%
发文量
2058
审稿时长
5.2 months
期刊介绍: The fields of interest of the IEEE Sensors Journal are the theory, design , fabrication, manufacturing and applications of devices for sensing and transducing physical, chemical and biological phenomena, with emphasis on the electronics and physics aspect of sensors and integrated sensors-actuators. IEEE Sensors Journal deals with the following: -Sensor Phenomenology, Modelling, and Evaluation -Sensor Materials, Processing, and Fabrication -Chemical and Gas Sensors -Microfluidics and Biosensors -Optical Sensors -Physical Sensors: Temperature, Mechanical, Magnetic, and others -Acoustic and Ultrasonic Sensors -Sensor Packaging -Sensor Networks -Sensor Applications -Sensor Systems: Signals, Processing, and Interfaces -Actuators and Sensor Power Systems -Sensor Signal Processing for high precision and stability (amplification, filtering, linearization, modulation/demodulation) and under harsh conditions (EMC, radiation, humidity, temperature); energy consumption/harvesting -Sensor Data Processing (soft computing with sensor data, e.g., pattern recognition, machine learning, evolutionary computation; sensor data fusion, processing of wave e.g., electromagnetic and acoustic; and non-wave, e.g., chemical, gravity, particle, thermal, radiative and non-radiative sensor data, detection, estimation and classification based on sensor data) -Sensors in Industrial Practice
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信