Automatic Image Recognition Meal Reporting Among Young Adults: Randomized Controlled Trial.

IF 6.2 2区 医学 Q1 HEALTH CARE SCIENCES & SERVICES
Prasan Kumar Sahoo, Sherry Yueh-Hsia Chiu, Yu-Sheng Lin, Chien-Hung Chen, Denisa Irianti, Hsin-Yun Chen, Mekhla Sarkar, Ying-Chieh Liu
{"title":"Automatic Image Recognition Meal Reporting Among Young Adults: Randomized Controlled Trial.","authors":"Prasan Kumar Sahoo, Sherry Yueh-Hsia Chiu, Yu-Sheng Lin, Chien-Hung Chen, Denisa Irianti, Hsin-Yun Chen, Mekhla Sarkar, Ying-Chieh Liu","doi":"10.2196/60070","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Advances in artificial intelligence technology have raised new possibilities for the effective evaluation of daily dietary intake, but more empirical study is needed for the use of such technologies under realistic meal scenarios. This study developed an automated food recognition technology, which was then integrated into its previous design to improve usability for meal reporting. The newly developed app allowed for the automatic detection and recognition of multiple dishes within a single real-time food image as input. App performance was tested using young adults in authentic dining conditions.</p><p><strong>Objective: </strong>A 2-group comparative study was conducted to assess app performance using metrics including accuracy, efficiency, and user perception. The experimental group, named the automatic image-based reporting (AIR) group, was compared against a control group using the previous version, named the voice input reporting (VIR) group. Each application is primarily designed to facilitate a distinct method of food intake reporting. AIR users capture and upload images of their selected dishes, supplemented with voice commands where appropriate. VIR users supplement the uploaded image with verbal inputs for food names and attributes.</p><p><strong>Methods: </strong>The 2 mobile apps were subjected to a head-to-head parallel randomized evaluation. A cohort of 42 young adults aged 20-25 years (9 male and 33 female participants) was recruited from a university in Taiwan and randomly assigned to 2 groups, that is, AIR (n=22) and VIR (n=20). Both groups were assessed using the same menu of 17 dishes. Each meal was designed to represent a typical lunch or dinner setting, with 1 staple, 1 main course, and 3 side dishes. All participants used the app on the same type of smartphone, with the interfaces of both using uniform user interactions, icons, and layouts. Analysis of the gathered data focused on assessing reporting accuracy, time efficiency, and user perception.</p><p><strong>Results: </strong>For the AIR group, 86% (189/220) of dishes were correctly identified, whereas 68% (136/200) of dishes were accurately reported. The AIR group exhibited a significantly higher degree of identification accuracy compared to the VIR group (P<.001). The AIR group also required significantly less time to complete food reporting (P<.001). System usability scale scores showed both apps were perceived as having high usability and learnability (P=.20).</p><p><strong>Conclusions: </strong>The AIR group outperformed the VIR group concerning accuracy and time efficiency for overall dish reporting within the meal testing scenario. While further technological enhancement may be required, artificial intelligence vision technology integration into existing mobile apps holds promise. Our results provide evidence-based contributions to the integration of automatic image recognition technology into existing apps in terms of user interaction efficacy and overall ease of use. Further empirical work is required, including full-scale randomized controlled trials and assessments of user perception under various conditions.</p>","PeriodicalId":14756,"journal":{"name":"JMIR mHealth and uHealth","volume":"13 ","pages":"e60070"},"PeriodicalIF":6.2000,"publicationDate":"2025-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12352700/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"JMIR mHealth and uHealth","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.2196/60070","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"HEALTH CARE SCIENCES & SERVICES","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Advances in artificial intelligence technology have raised new possibilities for the effective evaluation of daily dietary intake, but more empirical study is needed for the use of such technologies under realistic meal scenarios. This study developed an automated food recognition technology, which was then integrated into its previous design to improve usability for meal reporting. The newly developed app allowed for the automatic detection and recognition of multiple dishes within a single real-time food image as input. App performance was tested using young adults in authentic dining conditions.

Objective: A 2-group comparative study was conducted to assess app performance using metrics including accuracy, efficiency, and user perception. The experimental group, named the automatic image-based reporting (AIR) group, was compared against a control group using the previous version, named the voice input reporting (VIR) group. Each application is primarily designed to facilitate a distinct method of food intake reporting. AIR users capture and upload images of their selected dishes, supplemented with voice commands where appropriate. VIR users supplement the uploaded image with verbal inputs for food names and attributes.

Methods: The 2 mobile apps were subjected to a head-to-head parallel randomized evaluation. A cohort of 42 young adults aged 20-25 years (9 male and 33 female participants) was recruited from a university in Taiwan and randomly assigned to 2 groups, that is, AIR (n=22) and VIR (n=20). Both groups were assessed using the same menu of 17 dishes. Each meal was designed to represent a typical lunch or dinner setting, with 1 staple, 1 main course, and 3 side dishes. All participants used the app on the same type of smartphone, with the interfaces of both using uniform user interactions, icons, and layouts. Analysis of the gathered data focused on assessing reporting accuracy, time efficiency, and user perception.

Results: For the AIR group, 86% (189/220) of dishes were correctly identified, whereas 68% (136/200) of dishes were accurately reported. The AIR group exhibited a significantly higher degree of identification accuracy compared to the VIR group (P<.001). The AIR group also required significantly less time to complete food reporting (P<.001). System usability scale scores showed both apps were perceived as having high usability and learnability (P=.20).

Conclusions: The AIR group outperformed the VIR group concerning accuracy and time efficiency for overall dish reporting within the meal testing scenario. While further technological enhancement may be required, artificial intelligence vision technology integration into existing mobile apps holds promise. Our results provide evidence-based contributions to the integration of automatic image recognition technology into existing apps in terms of user interaction efficacy and overall ease of use. Further empirical work is required, including full-scale randomized controlled trials and assessments of user perception under various conditions.

在年轻人中自动图像识别进餐报告:随机对照试验。
背景:人工智能技术的进步为每日膳食摄入量的有效评估提供了新的可能性,但在现实膳食场景下使用人工智能技术需要更多的实证研究。本研究开发了一种自动食物识别技术,然后将其集成到先前的设计中,以提高膳食报告的可用性。这款新开发的应用程序可以在一张实时食物图像中自动检测和识别多个菜肴。应用程序的性能测试是在真实的用餐条件下使用年轻人进行的。目的:通过准确性、效率和用户感知等指标对应用性能进行两组对比研究。实验组被命名为自动图像报告(AIR)组,与使用先前版本的对照组被命名为语音输入报告(VIR)组进行比较。每个应用程序的设计主要是为了方便一种独特的食物摄入报告方法。AIR用户捕捉并上传他们所选菜肴的图片,并在适当的时候辅以语音命令。VIR用户在上传的图片上补充了食物名称和属性的口头输入。方法:对两款手机应用进行平行随机评价。本研究从台湾某大学招募了42名年龄在20-25岁的年轻人(男9人,女33人),随机分为AIR组(n=22)和VIR组(n=20)。两组人都使用相同的17道菜菜单进行评估。每餐都被设计成典型的午餐或晚餐,有1个主食,1个主菜和3个配菜。所有参与者在同一类型的智能手机上使用该应用程序,两者的界面使用统一的用户交互、图标和布局。对收集到的数据进行分析,重点是评估报告的准确性、时间效率和用户感知。结果:AIR组有86%(189/220)的菜品被正确识别,68%(136/200)的菜品被准确报告。与VIR组相比,AIR组表现出明显更高的识别准确性(结论:在膳食测试场景中,AIR组在总体菜肴报告的准确性和时间效率方面优于VIR组。虽然可能需要进一步的技术改进,但将人工智能视觉技术集成到现有的移动应用程序中是有希望的。我们的研究结果为将自动图像识别技术集成到现有应用程序中的用户交互效率和整体易用性提供了基于证据的贡献。需要进一步的实证工作,包括全面的随机对照试验和各种条件下用户感知的评估。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
JMIR mHealth and uHealth
JMIR mHealth and uHealth Medicine-Health Informatics
CiteScore
12.60
自引率
4.00%
发文量
159
审稿时长
10 weeks
期刊介绍: JMIR mHealth and uHealth (JMU, ISSN 2291-5222) is a spin-off journal of JMIR, the leading eHealth journal (Impact Factor 2016: 5.175). JMIR mHealth and uHealth is indexed in PubMed, PubMed Central, and Science Citation Index Expanded (SCIE), and in June 2017 received a stunning inaugural Impact Factor of 4.636. The journal focusses on health and biomedical applications in mobile and tablet computing, pervasive and ubiquitous computing, wearable computing and domotics. JMIR mHealth and uHealth publishes since 2013 and was the first mhealth journal in Pubmed. It publishes even faster and has a broader scope with including papers which are more technical or more formative/developmental than what would be published in the Journal of Medical Internet Research.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信