iCatcher+:从实验室、现场和在线研究中收集的视频中对婴幼儿凝视行为进行稳健和自动的注释。

IF 15.6 1区 心理学 Q1 PSYCHOLOGY
Yotam Erel, Katherine Adams Shannon, Junyi Chu, Kim Scott, Melissa Kline Struhl, Peng Cao, Xincheng Tan, Peter Hart, Gal Raz, Sabrina Piccolo, Catherine Mei, Christine Potter, Sagi Jaffe-Dax, Casey Lew-Williams, Joshua Tenenbaum, Katherine Fairchild, Amit Bermano, Shari Liu
{"title":"iCatcher+:从实验室、现场和在线研究中收集的视频中对婴幼儿凝视行为进行稳健和自动的注释。","authors":"Yotam Erel,&nbsp;Katherine Adams Shannon,&nbsp;Junyi Chu,&nbsp;Kim Scott,&nbsp;Melissa Kline Struhl,&nbsp;Peng Cao,&nbsp;Xincheng Tan,&nbsp;Peter Hart,&nbsp;Gal Raz,&nbsp;Sabrina Piccolo,&nbsp;Catherine Mei,&nbsp;Christine Potter,&nbsp;Sagi Jaffe-Dax,&nbsp;Casey Lew-Williams,&nbsp;Joshua Tenenbaum,&nbsp;Katherine Fairchild,&nbsp;Amit Bermano,&nbsp;Shari Liu","doi":"10.1177/25152459221147250","DOIUrl":null,"url":null,"abstract":"<p><p>Technological advances in psychological research have enabled large-scale studies of human behavior and streamlined pipelines for automatic processing of data. However, studies of infants and children have not fully reaped these benefits because the behaviors of interest, such as gaze duration and direction, still have to be extracted from video through a laborious process of manual annotation, even when these data are collected online. Recent advances in computer vision raise the possibility of automated annotation of these video data. In this article, we built on a system for automatic gaze annotation in young children, iCatcher, by engineering improvements and then training and testing the system (referred to hereafter as iCatcher+) on three data sets with substantial video and participant variability (214 videos collected in U.S. lab and field sites, 143 videos collected in Senegal field sites, and 265 videos collected via webcams in homes; participant age range = 4 months-3.5 years). When trained on each of these data sets, iCatcher+ performed with near human-level accuracy on held-out videos on distinguishing \"LEFT\" versus \"RIGHT\" and \"ON\" versus \"OFF\" looking behavior across all data sets. This high performance was achieved at the level of individual frames, experimental trials, and study videos; held across participant demographics (e.g., age, race/ethnicity), participant behavior (e.g., movement, head position), and video characteristics (e.g., luminance); and generalized to a fourth, entirely held-out online data set. We close by discussing next steps required to fully automate the life cycle of online infant and child behavioral studies, representing a key step toward enabling robust and high-throughput developmental research.</p>","PeriodicalId":55645,"journal":{"name":"Advances in Methods and Practices in Psychological Science","volume":"6 2","pages":""},"PeriodicalIF":15.6000,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ftp.ncbi.nlm.nih.gov/pub/pmc/oa_pdf/55/6e/nihms-1916587.PMC10471135.pdf","citationCount":"0","resultStr":"{\"title\":\"iCatcher+: Robust and Automated Annotation of Infants' and Young Children's Gaze Behavior From Videos Collected in Laboratory, Field, and Online Studies.\",\"authors\":\"Yotam Erel,&nbsp;Katherine Adams Shannon,&nbsp;Junyi Chu,&nbsp;Kim Scott,&nbsp;Melissa Kline Struhl,&nbsp;Peng Cao,&nbsp;Xincheng Tan,&nbsp;Peter Hart,&nbsp;Gal Raz,&nbsp;Sabrina Piccolo,&nbsp;Catherine Mei,&nbsp;Christine Potter,&nbsp;Sagi Jaffe-Dax,&nbsp;Casey Lew-Williams,&nbsp;Joshua Tenenbaum,&nbsp;Katherine Fairchild,&nbsp;Amit Bermano,&nbsp;Shari Liu\",\"doi\":\"10.1177/25152459221147250\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Technological advances in psychological research have enabled large-scale studies of human behavior and streamlined pipelines for automatic processing of data. However, studies of infants and children have not fully reaped these benefits because the behaviors of interest, such as gaze duration and direction, still have to be extracted from video through a laborious process of manual annotation, even when these data are collected online. Recent advances in computer vision raise the possibility of automated annotation of these video data. In this article, we built on a system for automatic gaze annotation in young children, iCatcher, by engineering improvements and then training and testing the system (referred to hereafter as iCatcher+) on three data sets with substantial video and participant variability (214 videos collected in U.S. lab and field sites, 143 videos collected in Senegal field sites, and 265 videos collected via webcams in homes; participant age range = 4 months-3.5 years). When trained on each of these data sets, iCatcher+ performed with near human-level accuracy on held-out videos on distinguishing \\\"LEFT\\\" versus \\\"RIGHT\\\" and \\\"ON\\\" versus \\\"OFF\\\" looking behavior across all data sets. This high performance was achieved at the level of individual frames, experimental trials, and study videos; held across participant demographics (e.g., age, race/ethnicity), participant behavior (e.g., movement, head position), and video characteristics (e.g., luminance); and generalized to a fourth, entirely held-out online data set. We close by discussing next steps required to fully automate the life cycle of online infant and child behavioral studies, representing a key step toward enabling robust and high-throughput developmental research.</p>\",\"PeriodicalId\":55645,\"journal\":{\"name\":\"Advances in Methods and Practices in Psychological Science\",\"volume\":\"6 2\",\"pages\":\"\"},\"PeriodicalIF\":15.6000,\"publicationDate\":\"2023-04-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://ftp.ncbi.nlm.nih.gov/pub/pmc/oa_pdf/55/6e/nihms-1916587.PMC10471135.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Advances in Methods and Practices in Psychological Science\",\"FirstCategoryId\":\"102\",\"ListUrlMain\":\"https://doi.org/10.1177/25152459221147250\",\"RegionNum\":1,\"RegionCategory\":\"心理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2023/4/18 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q1\",\"JCRName\":\"PSYCHOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advances in Methods and Practices in Psychological Science","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.1177/25152459221147250","RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2023/4/18 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"PSYCHOLOGY","Score":null,"Total":0}
引用次数: 0

摘要

心理学研究的技术进步使人们能够对人类行为进行大规模研究,并简化了数据自动处理的管道。然而,对婴儿和儿童的研究并没有完全获得这些好处,因为感兴趣的行为,如凝视持续时间和方向,仍然需要通过手动注释的费力过程从视频中提取,即使这些数据是在网上收集的。计算机视觉的最新进展提高了对这些视频数据进行自动注释的可能性。在本文中,我们构建了一个用于幼儿凝视自动注释的系统iCatcher,通过工程改进,然后在具有显著视频和参与者可变性的三个数据集上对系统(以下简称为iCatcher+)进行培训和测试(214个视频在美国实验室和现场采集,143个视频在塞内加尔现场采集,265个视频通过家庭网络摄像头采集;参与者年龄范围=4个月-3.5岁)。当在这些数据集上进行训练时,iCatcher+在所有数据集上以接近人类水平的准确度对手持视频进行了区分“左”与“右”以及“开”与“关”的行为。这种高性能是在单个帧、实验试验和研究视频的水平上实现的;在参与者人口统计数据(例如,年龄、种族/民族)、参与者行为(例如,运动、头部位置)和视频特征(例如,亮度)中进行;并推广到第四个完全公开的在线数据集。最后,我们讨论了完全自动化在线婴儿和儿童行为研究生命周期所需的下一步,这是实现稳健和高通量发展研究的关键一步。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

iCatcher+: Robust and Automated Annotation of Infants' and Young Children's Gaze Behavior From Videos Collected in Laboratory, Field, and Online Studies.

iCatcher+: Robust and Automated Annotation of Infants' and Young Children's Gaze Behavior From Videos Collected in Laboratory, Field, and Online Studies.

iCatcher+: Robust and Automated Annotation of Infants' and Young Children's Gaze Behavior From Videos Collected in Laboratory, Field, and Online Studies.

iCatcher+: Robust and Automated Annotation of Infants' and Young Children's Gaze Behavior From Videos Collected in Laboratory, Field, and Online Studies.

Technological advances in psychological research have enabled large-scale studies of human behavior and streamlined pipelines for automatic processing of data. However, studies of infants and children have not fully reaped these benefits because the behaviors of interest, such as gaze duration and direction, still have to be extracted from video through a laborious process of manual annotation, even when these data are collected online. Recent advances in computer vision raise the possibility of automated annotation of these video data. In this article, we built on a system for automatic gaze annotation in young children, iCatcher, by engineering improvements and then training and testing the system (referred to hereafter as iCatcher+) on three data sets with substantial video and participant variability (214 videos collected in U.S. lab and field sites, 143 videos collected in Senegal field sites, and 265 videos collected via webcams in homes; participant age range = 4 months-3.5 years). When trained on each of these data sets, iCatcher+ performed with near human-level accuracy on held-out videos on distinguishing "LEFT" versus "RIGHT" and "ON" versus "OFF" looking behavior across all data sets. This high performance was achieved at the level of individual frames, experimental trials, and study videos; held across participant demographics (e.g., age, race/ethnicity), participant behavior (e.g., movement, head position), and video characteristics (e.g., luminance); and generalized to a fourth, entirely held-out online data set. We close by discussing next steps required to fully automate the life cycle of online infant and child behavioral studies, representing a key step toward enabling robust and high-throughput developmental research.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
21.20
自引率
0.70%
发文量
16
期刊介绍: In 2021, Advances in Methods and Practices in Psychological Science will undergo a transition to become an open access journal. This journal focuses on publishing innovative developments in research methods, practices, and conduct within the field of psychological science. It embraces a wide range of areas and topics and encourages the integration of methodological and analytical questions. The aim of AMPPS is to bring the latest methodological advances to researchers from various disciplines, even those who are not methodological experts. Therefore, the journal seeks submissions that are accessible to readers with different research interests and that represent the diverse research trends within the field of psychological science. The types of content that AMPPS welcomes include articles that communicate advancements in methods, practices, and metascience, as well as empirical scientific best practices. Additionally, tutorials, commentaries, and simulation studies on new techniques and research tools are encouraged. The journal also aims to publish papers that bring advances from specialized subfields to a broader audience. Lastly, AMPPS accepts Registered Replication Reports, which focus on replicating important findings from previously published studies. Overall, the transition of Advances in Methods and Practices in Psychological Science to an open access journal aims to increase accessibility and promote the dissemination of new developments in research methods and practices within the field of psychological science.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信