{"title":"基于环境背景和人体的情绪识别","authors":"Cheng-Shan Jiang, Z. Liu","doi":"10.1109/ICPS58381.2023.10128084","DOIUrl":null,"url":null,"abstract":"To promote the humanized interactive experience of the intelligent device and system, emotional intelligence has become a popular research field in the human-machine interaction. The previous research on emotion recognition based on computer vision has mostly been carried out by analysing facial expression or body posture, and psychological studies show that scene context also contributes some important information on emotion recognition. In addition, most context-aware emotion recognition studies focus on exploring the relevance analysis of environmental semantics, but the influence of feature encoder on semantic information embedding has not been fully discussed. In this paper, we proposed a Global Semantic Feature Enhancement-Dual Stream Densely Connected Network (GSFE-DSDCN) to enhance global semantic information learning from the perspectives of dimension and spatial. Densely connected pattern is introduced to concatenate the shallow and deep layers output, which fuses the semantic information of low-dimensional geometric features and high-dimensional abstract context features together. The Global Multi-Scale Feature Recalibration (GMSFR) module expands the receptive field in spatial, which effectively improves the global semantic features extraction capability of feature encoder. We evaluate the proposed method on the EMOTIC data set, and experimental results are shown to be competitive with the state-of-the-art algorithms.","PeriodicalId":426122,"journal":{"name":"2023 IEEE 6th International Conference on Industrial Cyber-Physical Systems (ICPS)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Emotion Recognition via Environmental Context and Human Body\",\"authors\":\"Cheng-Shan Jiang, Z. Liu\",\"doi\":\"10.1109/ICPS58381.2023.10128084\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"To promote the humanized interactive experience of the intelligent device and system, emotional intelligence has become a popular research field in the human-machine interaction. The previous research on emotion recognition based on computer vision has mostly been carried out by analysing facial expression or body posture, and psychological studies show that scene context also contributes some important information on emotion recognition. In addition, most context-aware emotion recognition studies focus on exploring the relevance analysis of environmental semantics, but the influence of feature encoder on semantic information embedding has not been fully discussed. In this paper, we proposed a Global Semantic Feature Enhancement-Dual Stream Densely Connected Network (GSFE-DSDCN) to enhance global semantic information learning from the perspectives of dimension and spatial. Densely connected pattern is introduced to concatenate the shallow and deep layers output, which fuses the semantic information of low-dimensional geometric features and high-dimensional abstract context features together. The Global Multi-Scale Feature Recalibration (GMSFR) module expands the receptive field in spatial, which effectively improves the global semantic features extraction capability of feature encoder. We evaluate the proposed method on the EMOTIC data set, and experimental results are shown to be competitive with the state-of-the-art algorithms.\",\"PeriodicalId\":426122,\"journal\":{\"name\":\"2023 IEEE 6th International Conference on Industrial Cyber-Physical Systems (ICPS)\",\"volume\":\"39 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-05-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 IEEE 6th International Conference on Industrial Cyber-Physical Systems (ICPS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICPS58381.2023.10128084\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 6th International Conference on Industrial Cyber-Physical Systems (ICPS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICPS58381.2023.10128084","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Emotion Recognition via Environmental Context and Human Body
To promote the humanized interactive experience of the intelligent device and system, emotional intelligence has become a popular research field in the human-machine interaction. The previous research on emotion recognition based on computer vision has mostly been carried out by analysing facial expression or body posture, and psychological studies show that scene context also contributes some important information on emotion recognition. In addition, most context-aware emotion recognition studies focus on exploring the relevance analysis of environmental semantics, but the influence of feature encoder on semantic information embedding has not been fully discussed. In this paper, we proposed a Global Semantic Feature Enhancement-Dual Stream Densely Connected Network (GSFE-DSDCN) to enhance global semantic information learning from the perspectives of dimension and spatial. Densely connected pattern is introduced to concatenate the shallow and deep layers output, which fuses the semantic information of low-dimensional geometric features and high-dimensional abstract context features together. The Global Multi-Scale Feature Recalibration (GMSFR) module expands the receptive field in spatial, which effectively improves the global semantic features extraction capability of feature encoder. We evaluate the proposed method on the EMOTIC data set, and experimental results are shown to be competitive with the state-of-the-art algorithms.