Kelly G Garner, Christopher R Nolan, Abbey Nydam, Zoie Nott, Howard Bowman, Paul E Dux
{"title":"量化注意力、执行功能和内隐学习中效应大小估计的误差。","authors":"Kelly G Garner, Christopher R Nolan, Abbey Nydam, Zoie Nott, Howard Bowman, Paul E Dux","doi":"10.1037/xlm0001338","DOIUrl":null,"url":null,"abstract":"<p><p>Accurate quantification of effect sizes has the power to motivate theory and reduce misinvestment of scientific resources by informing power calculations during study planning. However, a combination of publication bias and small sample sizes (∼<i>N</i> = 25) hampers certainty in current effect size estimates. We sought to determine the extent to which sample sizes may produce errors in effect size estimates for four commonly used paradigms assessing attention, executive function, and implicit learning (attentional blink, multitasking, contextual cueing, and serial response task). We combined a large data set with a bootstrapping approach to simulate 1,000 experiments across a range of N (13-313). Beyond quantifying the effect size and statistical power that can be anticipated for each study design, we demonstrate that experiments with lower N may double or triple information loss. We also show that basing power calculations on effect sizes from similar studies yields a problematically imprecise estimate between 40% and 67% of the time, given commonly used sample sizes. Last, we show that skewness of intersubject behavioral effects may serve as a predictor of an erroneous estimate. We conclude with practical recommendations for researchers and demonstrate how our simulation approach can yield theoretical insights that are not readily achieved by other methods such as identifying the information gained from rejecting the null hypothesis and quantifying the contribution of individual variation to error in effect size estimates. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":50194,"journal":{"name":"Journal of Experimental Psychology-Learning Memory and Cognition","volume":null,"pages":null},"PeriodicalIF":2.2000,"publicationDate":"2024-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Quantifying error in effect size estimates in attention, executive function, and implicit learning.\",\"authors\":\"Kelly G Garner, Christopher R Nolan, Abbey Nydam, Zoie Nott, Howard Bowman, Paul E Dux\",\"doi\":\"10.1037/xlm0001338\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Accurate quantification of effect sizes has the power to motivate theory and reduce misinvestment of scientific resources by informing power calculations during study planning. However, a combination of publication bias and small sample sizes (∼<i>N</i> = 25) hampers certainty in current effect size estimates. We sought to determine the extent to which sample sizes may produce errors in effect size estimates for four commonly used paradigms assessing attention, executive function, and implicit learning (attentional blink, multitasking, contextual cueing, and serial response task). We combined a large data set with a bootstrapping approach to simulate 1,000 experiments across a range of N (13-313). Beyond quantifying the effect size and statistical power that can be anticipated for each study design, we demonstrate that experiments with lower N may double or triple information loss. We also show that basing power calculations on effect sizes from similar studies yields a problematically imprecise estimate between 40% and 67% of the time, given commonly used sample sizes. Last, we show that skewness of intersubject behavioral effects may serve as a predictor of an erroneous estimate. We conclude with practical recommendations for researchers and demonstrate how our simulation approach can yield theoretical insights that are not readily achieved by other methods such as identifying the information gained from rejecting the null hypothesis and quantifying the contribution of individual variation to error in effect size estimates. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>\",\"PeriodicalId\":50194,\"journal\":{\"name\":\"Journal of Experimental Psychology-Learning Memory and Cognition\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":2.2000,\"publicationDate\":\"2024-07-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Experimental Psychology-Learning Memory and Cognition\",\"FirstCategoryId\":\"102\",\"ListUrlMain\":\"https://doi.org/10.1037/xlm0001338\",\"RegionNum\":2,\"RegionCategory\":\"心理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"PSYCHOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Experimental Psychology-Learning Memory and Cognition","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.1037/xlm0001338","RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"PSYCHOLOGY","Score":null,"Total":0}
引用次数: 0
摘要
效应大小的精确量化可以为研究规划中的功率计算提供信息,从而推动理论研究,减少科学资源的错误投资。然而,发表偏倚和小样本量(∼N = 25)共同影响了目前效应大小估计的确定性。我们试图确定样本量在多大程度上会对评估注意力、执行功能和内隐学习的四种常用范式(注意力眨眼、多任务处理、情境提示和连续反应任务)的效应大小估计产生误差。我们将大型数据集与自引导方法相结合,模拟了 1,000 个 N(13-313)范围内的实验。除了量化每种研究设计可预期的效应大小和统计功率外,我们还证明了较低 N 的实验可能会使信息损失增加一倍或两倍。我们还表明,根据类似研究的效应大小来计算统计能力,在常用样本量的情况下,40% 到 67% 的时间会得出不精确的估计值,这是有问题的。最后,我们还表明,受试者间行为效应的偏度可以预测错误的估计值。最后,我们为研究人员提供了实用建议,并展示了我们的模拟方法如何产生其他方法难以达到的理论洞察力,例如识别拒绝零假设所获得的信息,以及量化个体差异对效应大小估计误差的贡献。(PsycInfo Database Record (c) 2024 APA, 版权所有)。
Quantifying error in effect size estimates in attention, executive function, and implicit learning.
Accurate quantification of effect sizes has the power to motivate theory and reduce misinvestment of scientific resources by informing power calculations during study planning. However, a combination of publication bias and small sample sizes (∼N = 25) hampers certainty in current effect size estimates. We sought to determine the extent to which sample sizes may produce errors in effect size estimates for four commonly used paradigms assessing attention, executive function, and implicit learning (attentional blink, multitasking, contextual cueing, and serial response task). We combined a large data set with a bootstrapping approach to simulate 1,000 experiments across a range of N (13-313). Beyond quantifying the effect size and statistical power that can be anticipated for each study design, we demonstrate that experiments with lower N may double or triple information loss. We also show that basing power calculations on effect sizes from similar studies yields a problematically imprecise estimate between 40% and 67% of the time, given commonly used sample sizes. Last, we show that skewness of intersubject behavioral effects may serve as a predictor of an erroneous estimate. We conclude with practical recommendations for researchers and demonstrate how our simulation approach can yield theoretical insights that are not readily achieved by other methods such as identifying the information gained from rejecting the null hypothesis and quantifying the contribution of individual variation to error in effect size estimates. (PsycInfo Database Record (c) 2024 APA, all rights reserved).
期刊介绍:
The Journal of Experimental Psychology: Learning, Memory, and Cognition publishes studies on perception, control of action, perceptual aspects of language processing, and related cognitive processes.