Facial Expression Recognition using Spatial Feature Extraction and Ensemble Deep Networks

E. Afshar, Hassan Khotanlou, Elham Alighardash
{"title":"Facial Expression Recognition using Spatial Feature Extraction and Ensemble Deep Networks","authors":"E. Afshar, Hassan Khotanlou, Elham Alighardash","doi":"10.1109/IPRIA59240.2023.10147196","DOIUrl":null,"url":null,"abstract":"Researchers have shown that 55% of concepts are conveyed through facial emotion and only 7% are conveyed by words and sentences, so facial expression plays an important role in conveying concepts in human communications. In recent years, due to the improvement of artificial neural networks, many studies have been conducted related to facial expression recognition. This paper presents a method based on ensemble classification using convolutional neural networks to recognize facial emotions. The concatenation of spatial features with global features is used as a feature map for the classification stage in the committee network. Two committee networks are fed separately with LBP and raw images. After training the two committee networks, to classify the emotion, the maximum probability between the two networks is considered as the final output. The proposed method was applied and tested on the FER2013 dataset. Our proposed method is more accurate than many leading methods, and in competition with the successful model that has a more complex architecture and higher computational cost, it has been able to achieve acceptable results with a simple architecture.","PeriodicalId":109390,"journal":{"name":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 6th International Conference on Pattern Recognition and Image Analysis (IPRIA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IPRIA59240.2023.10147196","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Researchers have shown that 55% of concepts are conveyed through facial emotion and only 7% are conveyed by words and sentences, so facial expression plays an important role in conveying concepts in human communications. In recent years, due to the improvement of artificial neural networks, many studies have been conducted related to facial expression recognition. This paper presents a method based on ensemble classification using convolutional neural networks to recognize facial emotions. The concatenation of spatial features with global features is used as a feature map for the classification stage in the committee network. Two committee networks are fed separately with LBP and raw images. After training the two committee networks, to classify the emotion, the maximum probability between the two networks is considered as the final output. The proposed method was applied and tested on the FER2013 dataset. Our proposed method is more accurate than many leading methods, and in competition with the successful model that has a more complex architecture and higher computational cost, it has been able to achieve acceptable results with a simple architecture.
基于空间特征提取和集成深度网络的面部表情识别
研究表明,55%的概念是通过面部表情传达的,只有7%的概念是通过文字和句子传达的,因此面部表情在人类交流中对概念的传达起着重要的作用。近年来,由于人工神经网络的改进,人们对面部表情识别进行了很多相关的研究。提出了一种基于集成分类的卷积神经网络人脸情绪识别方法。在委员会网络中,空间特征与全局特征的连接被用作分类阶段的特征映射。两个委员会网络分别输入LBP和原始图像。在对两个委员会网络进行训练后,对情感进行分类,将两个网络之间的最大概率作为最终输出。该方法在FER2013数据集上进行了应用和测试。我们提出的方法比许多领先的方法更精确,并且在与结构更复杂、计算成本更高的成功模型的竞争中,它已经能够以简单的结构获得可接受的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信