视频反应中捕捉到的面部表情分析

K. Srinivasa, Sriram Anupindi, R. Sharath, S. Chaitanya
{"title":"视频反应中捕捉到的面部表情分析","authors":"K. Srinivasa, Sriram Anupindi, R. Sharath, S. Chaitanya","doi":"10.1109/IACC.2017.0140","DOIUrl":null,"url":null,"abstract":"With the advent of technology, there has been a rapid increase in data acquisition. Among the several types of data gathered, interpreting multimedia data by a machine without human intervention is a challenge. Extracting meaningful content from videos will help provide better solutions in various domains. Banking on processing the videos as our rudimentary concept, this paper intends to detect the expressiveness of an individual. Many traditional approaches exist to address this situation, however Deep learning is used in this work to achieve the goal. LSTM (Long Short Term Memory) is our selected implementation construct. For training the network MIT Affectiva dataset has been chosen, which comprises of videos of individuals responding to Superbowl commercials. Extending this network, two front ends for testing the video sample are provided. One is a web page for uploading the videos and displaying the results and the other is an IoT device. This device records the video of an individual's response and sends it to the processing server. The responses of individuals towards a particular commercial are recorded and the system is tested on it. The results obtained are examined and the scope for various interpretations are shown. The expressiveness detected becomes a pivotal feedback for the makers of commercials, paving the way for their improvement.","PeriodicalId":248433,"journal":{"name":"2017 IEEE 7th International Advance Computing Conference (IACC)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"Analysis of Facial Expressiveness Captured in Reaction to Videos\",\"authors\":\"K. Srinivasa, Sriram Anupindi, R. Sharath, S. Chaitanya\",\"doi\":\"10.1109/IACC.2017.0140\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the advent of technology, there has been a rapid increase in data acquisition. Among the several types of data gathered, interpreting multimedia data by a machine without human intervention is a challenge. Extracting meaningful content from videos will help provide better solutions in various domains. Banking on processing the videos as our rudimentary concept, this paper intends to detect the expressiveness of an individual. Many traditional approaches exist to address this situation, however Deep learning is used in this work to achieve the goal. LSTM (Long Short Term Memory) is our selected implementation construct. For training the network MIT Affectiva dataset has been chosen, which comprises of videos of individuals responding to Superbowl commercials. Extending this network, two front ends for testing the video sample are provided. One is a web page for uploading the videos and displaying the results and the other is an IoT device. This device records the video of an individual's response and sends it to the processing server. The responses of individuals towards a particular commercial are recorded and the system is tested on it. The results obtained are examined and the scope for various interpretations are shown. The expressiveness detected becomes a pivotal feedback for the makers of commercials, paving the way for their improvement.\",\"PeriodicalId\":248433,\"journal\":{\"name\":\"2017 IEEE 7th International Advance Computing Conference (IACC)\",\"volume\":\"39 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1900-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2017 IEEE 7th International Advance Computing Conference (IACC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IACC.2017.0140\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 IEEE 7th International Advance Computing Conference (IACC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IACC.2017.0140","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

摘要

随着技术的发展,数据采集量迅速增加。在收集的几种类型的数据中,由机器在没有人为干预的情况下解释多媒体数据是一个挑战。从视频中提取有意义的内容将有助于在各个领域提供更好的解决方案。本文以处理视频作为我们的基本概念,旨在检测个体的表现力。有许多传统的方法可以解决这种情况,但是在这项工作中使用深度学习来实现目标。LSTM(长短期记忆)是我们选择的实现结构。为了训练网络,选择了麻省理工学院(MIT)的Affectiva数据集,该数据集包括个人对超级碗广告的反应视频。对该网络进行扩展,提供了用于测试视频样本的两个前端。一个是用于上传视频和显示结果的网页,另一个是物联网设备。这个设备记录下个人反应的视频并将其发送到处理服务器。个人对某一特定广告的反应会被记录下来,并对系统进行测试。对所得结果进行了检验,并说明了各种解释的范围。检测到的表达能力成为广告制作者的关键反馈,为他们的改进铺平道路。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Analysis of Facial Expressiveness Captured in Reaction to Videos
With the advent of technology, there has been a rapid increase in data acquisition. Among the several types of data gathered, interpreting multimedia data by a machine without human intervention is a challenge. Extracting meaningful content from videos will help provide better solutions in various domains. Banking on processing the videos as our rudimentary concept, this paper intends to detect the expressiveness of an individual. Many traditional approaches exist to address this situation, however Deep learning is used in this work to achieve the goal. LSTM (Long Short Term Memory) is our selected implementation construct. For training the network MIT Affectiva dataset has been chosen, which comprises of videos of individuals responding to Superbowl commercials. Extending this network, two front ends for testing the video sample are provided. One is a web page for uploading the videos and displaying the results and the other is an IoT device. This device records the video of an individual's response and sends it to the processing server. The responses of individuals towards a particular commercial are recorded and the system is tested on it. The results obtained are examined and the scope for various interpretations are shown. The expressiveness detected becomes a pivotal feedback for the makers of commercials, paving the way for their improvement.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信