Konstantin Warneke, Manuel Oraže, Gerit Plöschberger, Marco Herbsleb, Jose Afonso, Sebastian Wallot
{"title":"When Testing Becomes Learning—Underscoring the Relevance of Habituation to Improve Internal Validity of Common Neurocognitive Tests","authors":"Konstantin Warneke, Manuel Oraže, Gerit Plöschberger, Marco Herbsleb, Jose Afonso, Sebastian Wallot","doi":"10.1111/ejn.70117","DOIUrl":null,"url":null,"abstract":"<p>Testing neurocognitive function is receiving growing attention in psychological and physical health research. To counteract the costs, reduced accessibility, and complexity of brain imaging (e.g., CT scans and fMRI) or function tests, neurocognitive performance tests (e.g., the Stroop test, the Trail Making Test, or the Choice Reaction Task) are commonly implemented. Although reliability is considered paramount when interpreting intervention effects, a detailed quantification of systematic and random errors is scarce. By recruiting 68 healthy participants from different age groups (7–64 years), we quantified population-specific measurement errors in the aforementioned neurocognitive tasks. The goal was to raise awareness about the impact of learning effects on reliability assessments and their interpretation. By performing five testing sessions with two trials per day, we observed significant learning effects from repeated testing. Trial-to-trial improvements of up to 50% were measured, accompanied by a random measurement error reduction from day to day. These learning effects were task and population specific, highlighting the need for caution when transferring reliability coefficients from other studies. The quantification of systematic and random measurement errors underscores the importance of conducting sufficient habituation sessions in neurocognitive tasks, as test protocols lack validity if they do not ensure reliability. Therefore, sufficient habituation sessions (i.e., until no meaningful learning effects can be observed) may be warranted when testing is repeated within short timeframes.</p>","PeriodicalId":11993,"journal":{"name":"European Journal of Neuroscience","volume":"61 8","pages":""},"PeriodicalIF":2.4000,"publicationDate":"2025-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/ejn.70117","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"European Journal of Neuroscience","FirstCategoryId":"3","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/ejn.70117","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"NEUROSCIENCES","Score":null,"Total":0}
引用次数: 0
Abstract
Testing neurocognitive function is receiving growing attention in psychological and physical health research. To counteract the costs, reduced accessibility, and complexity of brain imaging (e.g., CT scans and fMRI) or function tests, neurocognitive performance tests (e.g., the Stroop test, the Trail Making Test, or the Choice Reaction Task) are commonly implemented. Although reliability is considered paramount when interpreting intervention effects, a detailed quantification of systematic and random errors is scarce. By recruiting 68 healthy participants from different age groups (7–64 years), we quantified population-specific measurement errors in the aforementioned neurocognitive tasks. The goal was to raise awareness about the impact of learning effects on reliability assessments and their interpretation. By performing five testing sessions with two trials per day, we observed significant learning effects from repeated testing. Trial-to-trial improvements of up to 50% were measured, accompanied by a random measurement error reduction from day to day. These learning effects were task and population specific, highlighting the need for caution when transferring reliability coefficients from other studies. The quantification of systematic and random measurement errors underscores the importance of conducting sufficient habituation sessions in neurocognitive tasks, as test protocols lack validity if they do not ensure reliability. Therefore, sufficient habituation sessions (i.e., until no meaningful learning effects can be observed) may be warranted when testing is repeated within short timeframes.
神经认知功能测试在心理和生理健康研究中受到越来越多的关注。为了抵消脑成像(如CT扫描和fMRI)或功能测试的成本、可及性降低和复杂性,通常实施神经认知性能测试(如Stroop测试、Trail Making test或Choice Reaction Task)。虽然在解释干预效果时,可靠性被认为是最重要的,但对系统和随机误差的详细量化却很少。通过招募来自不同年龄组(7-64岁)的68名健康参与者,我们量化了上述神经认知任务中特定人群的测量误差。目的是提高人们对学习效应对可靠性评估及其解释的影响的认识。通过进行5次测试,每天2次,我们观察到重复测试的显著学习效果。试验对试验的改善高达50%,伴随着每天随机测量误差的减少。这些学习效应是特定于任务和人群的,强调了在从其他研究中转移可靠性系数时需要谨慎。系统和随机测量误差的量化强调了在神经认知任务中进行足够的习惯化过程的重要性,因为如果测试方案不能确保可靠性,那么它们就缺乏有效性。因此,当测试在短时间内重复时,充分的习惯化过程(即,直到没有观察到有意义的学习效果为止)可能是有保证的。
期刊介绍:
EJN is the journal of FENS and supports the international neuroscientific community by publishing original high quality research articles and reviews in all fields of neuroscience. In addition, to engage with issues that are of interest to the science community, we also publish Editorials, Meetings Reports and Neuro-Opinions on topics that are of current interest in the fields of neuroscience research and training in science. We have recently established a series of ‘Profiles of Women in Neuroscience’. Our goal is to provide a vehicle for publications that further the understanding of the structure and function of the nervous system in both health and disease and to provide a vehicle to engage the neuroscience community. As the official journal of FENS, profits from the journal are re-invested in the neuroscientific community through the activities of FENS.