{"title":"AI-edge based voice responsive smart headphone for user context-awarenes","authors":"V. V, K. M, G. Reddy","doi":"10.1109/CONECCT50063.2020.9198484","DOIUrl":null,"url":null,"abstract":"Recent developments in the areas of portable devices have taken enormous developments, especially the audio/music information delivery systems like headphones either using wired or wireless connectivity with mobile/portable devices. Usage of such portable devices with connected headphones is increasing day by day with user mobility in almost all the human age groups and sectors like traveling, workplaces, home places, walk paths, etc. This usage has an issue with the user interaction with the external world and also the contextual situations, actions and responses. Such issues lead to non-involvement and inconvenience or total destruction to the user. This depends on the surrounding situation and contextual hazardous events that may lead to accidents or life-threatening injuries. In order to handle such issues, we have designed an edge-based AI Voice responsive smart headphone for context awareness and alert system to the user. This model uses an edge-based machine learning Long Short Term Memory (LSTM) algorithm for voice recognition, runs either in the headphone or the device to which it is connected. Here the voice/sound recognition and alerting the user, can handle many contextual situations like normal interactions by name, hazardous scary situational sounds either by a bystander, or external sounds. Prototype models are developed for some of the voice records/ sounds using tensor flow ML-based algorithms over low footprint devices like mobile phones or STM 32 microcontrollers. Tested and validated for some of the voice/sound scenarios.","PeriodicalId":261794,"journal":{"name":"2020 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CONECCT50063.2020.9198484","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Recent developments in the areas of portable devices have taken enormous developments, especially the audio/music information delivery systems like headphones either using wired or wireless connectivity with mobile/portable devices. Usage of such portable devices with connected headphones is increasing day by day with user mobility in almost all the human age groups and sectors like traveling, workplaces, home places, walk paths, etc. This usage has an issue with the user interaction with the external world and also the contextual situations, actions and responses. Such issues lead to non-involvement and inconvenience or total destruction to the user. This depends on the surrounding situation and contextual hazardous events that may lead to accidents or life-threatening injuries. In order to handle such issues, we have designed an edge-based AI Voice responsive smart headphone for context awareness and alert system to the user. This model uses an edge-based machine learning Long Short Term Memory (LSTM) algorithm for voice recognition, runs either in the headphone or the device to which it is connected. Here the voice/sound recognition and alerting the user, can handle many contextual situations like normal interactions by name, hazardous scary situational sounds either by a bystander, or external sounds. Prototype models are developed for some of the voice records/ sounds using tensor flow ML-based algorithms over low footprint devices like mobile phones or STM 32 microcontrollers. Tested and validated for some of the voice/sound scenarios.