A Resource Efficient System for On-Smartwatch Audio Processing.

Md Sabbir Ahmed, Arafat Rahman, Zhiyuan Wang, Mark Rucker, Laura E Barnes
{"title":"A Resource Efficient System for On-Smartwatch Audio Processing.","authors":"Md Sabbir Ahmed, Arafat Rahman, Zhiyuan Wang, Mark Rucker, Laura E Barnes","doi":"10.1145/3636534.3698866","DOIUrl":null,"url":null,"abstract":"<p><p>While audio data shows promise in addressing various health challenges, there is a lack of research on on-device audio processing for smartwatches. Privacy concerns make storing raw audio and performing post-hoc analysis undesirable for many users. Additionally, current on-device audio processing systems for smartwatches are limited in their feature extraction capabilities, restricting their potential for understanding user behavior and health. We developed a real-time system for on-device audio processing on smartwatches, which takes an average of 1.78 minutes (SD = 0.07 min) to extract 22 spectral and rhythmic features from a 1-minute audio sample, using a small window size of 25 milliseconds. Using these extracted audio features on a public dataset, we developed and incorporated models into a watch to classify foreground and background speech in real-time. Our Random Forest-based model classifies speech with a balanced accuracy of 80.3%.</p>","PeriodicalId":91382,"journal":{"name":"Proceedings of the ... annual International Conference on Mobile Computing and Networking. International Conference on Mobile Computing and Networking","volume":"2024 ","pages":"1805-1807"},"PeriodicalIF":0.0000,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12126283/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the ... annual International Conference on Mobile Computing and Networking. International Conference on Mobile Computing and Networking","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3636534.3698866","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/12/4 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

While audio data shows promise in addressing various health challenges, there is a lack of research on on-device audio processing for smartwatches. Privacy concerns make storing raw audio and performing post-hoc analysis undesirable for many users. Additionally, current on-device audio processing systems for smartwatches are limited in their feature extraction capabilities, restricting their potential for understanding user behavior and health. We developed a real-time system for on-device audio processing on smartwatches, which takes an average of 1.78 minutes (SD = 0.07 min) to extract 22 spectral and rhythmic features from a 1-minute audio sample, using a small window size of 25 milliseconds. Using these extracted audio features on a public dataset, we developed and incorporated models into a watch to classify foreground and background speech in real-time. Our Random Forest-based model classifies speech with a balanced accuracy of 80.3%.

智能手表音频处理的资源高效系统。
虽然音频数据有望解决各种健康挑战,但缺乏对智能手表设备上音频处理的研究。出于隐私考虑,许多用户不希望存储原始音频并执行事后分析。此外,目前用于智能手表的设备音频处理系统在特征提取能力方面受到限制,限制了它们理解用户行为和健康状况的潜力。我们开发了一个用于智能手表设备上音频处理的实时系统,平均需要1.78分钟(SD = 0.07分钟)从1分钟音频样本中提取22个频谱和节奏特征,使用25毫秒的小窗口大小。利用这些在公共数据集上提取的音频特征,我们开发并将模型整合到手表中,以实时分类前景和背景语音。我们基于随机森林的模型以80.3%的平衡准确率对语音进行分类。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信