Akshi:一个使用机器学习的视觉障碍辅助系统

Aakash Jain, Ritik Verma, Gurtej Singh Khokhar, Madhulika Bhadauria
{"title":"Akshi:一个使用机器学习的视觉障碍辅助系统","authors":"Aakash Jain, Ritik Verma, Gurtej Singh Khokhar, Madhulika Bhadauria","doi":"10.1109/AIST55798.2022.10064996","DOIUrl":null,"url":null,"abstract":"This work focuses on is emotion recognition. Emotion shows crucial data about human communication. It’s general to utilize face expressions to convey feelings throughout a discussion, and personal communication is only possible through facial expressions. The goal of this study is for offering a machine learning-based emotion recognition structure for people who are impaired visually. We present a CNN-based solution approach to manage this challenge, for training and testing we used FER2013 database which consisted of 7 facial expression and a total of 35,685 images out of which we selected 3 facial expression happy, sad, and neutral comprising of 21264 images and achieved an accuracy of 81%.It has some limitations that it needs a person to operate and sometimes mix up of expressions so gives wrong results. Likewise, CNN we also implemented Transfer learning model i.e., Mobile Net for facial expression detection with similar dataset and achieved an accuracy of around 80%. To improvise our overall accuracy firstly we and modified the CNN design and achieved an overall accuracy of 91.65% which was superior to previous two implementations. The primary utility of the model is to help visually impaired people for better communication.","PeriodicalId":360351,"journal":{"name":"2022 4th International Conference on Artificial Intelligence and Speech Technology (AIST)","volume":"156 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Akshi: An Assistance system for visually challenged using Machine Learning\",\"authors\":\"Aakash Jain, Ritik Verma, Gurtej Singh Khokhar, Madhulika Bhadauria\",\"doi\":\"10.1109/AIST55798.2022.10064996\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This work focuses on is emotion recognition. Emotion shows crucial data about human communication. It’s general to utilize face expressions to convey feelings throughout a discussion, and personal communication is only possible through facial expressions. The goal of this study is for offering a machine learning-based emotion recognition structure for people who are impaired visually. We present a CNN-based solution approach to manage this challenge, for training and testing we used FER2013 database which consisted of 7 facial expression and a total of 35,685 images out of which we selected 3 facial expression happy, sad, and neutral comprising of 21264 images and achieved an accuracy of 81%.It has some limitations that it needs a person to operate and sometimes mix up of expressions so gives wrong results. Likewise, CNN we also implemented Transfer learning model i.e., Mobile Net for facial expression detection with similar dataset and achieved an accuracy of around 80%. To improvise our overall accuracy firstly we and modified the CNN design and achieved an overall accuracy of 91.65% which was superior to previous two implementations. The primary utility of the model is to help visually impaired people for better communication.\",\"PeriodicalId\":360351,\"journal\":{\"name\":\"2022 4th International Conference on Artificial Intelligence and Speech Technology (AIST)\",\"volume\":\"156 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-12-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 4th International Conference on Artificial Intelligence and Speech Technology (AIST)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/AIST55798.2022.10064996\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 4th International Conference on Artificial Intelligence and Speech Technology (AIST)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AIST55798.2022.10064996","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

这项工作的重点是情绪识别。情感显示了人类交流的关键数据。在整个讨论过程中,通常使用面部表情来传达情感,而个人交流只能通过面部表情来实现。本研究的目的是为视障人士提供一种基于机器学习的情感识别结构。我们提出了一种基于cnn的解决方案来应对这一挑战,对于训练和测试,我们使用FER2013数据库,该数据库由7个面部表情组成,共35,685张图像,其中我们选择了3个面部表情,包括21264张图像,包括快乐,悲伤和中性,达到了81%的准确率。它有一些局限性,需要一个人来操作,有时会混淆表达,从而给出错误的结果。同样,CNN我们也实现了迁移学习模型,即Mobile Net,使用相似的数据集进行面部表情检测,准确率达到80%左右。为了提高我们的整体精度,我们首先修改了CNN的设计,实现了91.65%的整体精度,优于前两次实现。该模型的主要用途是帮助视障人士更好地沟通。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Akshi: An Assistance system for visually challenged using Machine Learning
This work focuses on is emotion recognition. Emotion shows crucial data about human communication. It’s general to utilize face expressions to convey feelings throughout a discussion, and personal communication is only possible through facial expressions. The goal of this study is for offering a machine learning-based emotion recognition structure for people who are impaired visually. We present a CNN-based solution approach to manage this challenge, for training and testing we used FER2013 database which consisted of 7 facial expression and a total of 35,685 images out of which we selected 3 facial expression happy, sad, and neutral comprising of 21264 images and achieved an accuracy of 81%.It has some limitations that it needs a person to operate and sometimes mix up of expressions so gives wrong results. Likewise, CNN we also implemented Transfer learning model i.e., Mobile Net for facial expression detection with similar dataset and achieved an accuracy of around 80%. To improvise our overall accuracy firstly we and modified the CNN design and achieved an overall accuracy of 91.65% which was superior to previous two implementations. The primary utility of the model is to help visually impaired people for better communication.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信