Jan Kadlec, Catherine R. Walsh, Uri Sadé, Ariel Amir, Jesse Rissman, Michal Ramot
{"title":"A measure of reliability convergence to select and optimize cognitive tasks for individual differences research","authors":"Jan Kadlec, Catherine R. Walsh, Uri Sadé, Ariel Amir, Jesse Rissman, Michal Ramot","doi":"10.1038/s44271-024-00114-4","DOIUrl":null,"url":null,"abstract":"Surging interest in individual differences has faced setbacks in light of recent replication crises in psychology, for example in brain-wide association studies exploring brain-behavior correlations. A crucial component of replicability for individual differences studies, which is often assumed but not directly tested, is the reliability of the measures we use. Here, we evaluate the reliability of different cognitive tasks on a dataset with over 250 participants, who each completed a multi-day task battery. We show how reliability improves as a function of number of trials, and describe the convergence of the reliability curves for the different tasks, allowing us to score tasks according to their suitability for studies of individual differences. We further show the effect on reliability of measuring over multiple time points, with tasks assessing different cognitive domains being differentially affected. Data collected over more than one session may be required to achieve trait-like stability. Reliability of cognitive task measures improves as a function of number of trials. Because of differences in reliability convergence, tasks differ in suitability as estimates of individual differences. To achieve traitlike stability in measures, data must be combined across sessions.","PeriodicalId":501698,"journal":{"name":"Communications Psychology","volume":" ","pages":"1-18"},"PeriodicalIF":0.0000,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.com/articles/s44271-024-00114-4.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Communications Psychology","FirstCategoryId":"1085","ListUrlMain":"https://www.nature.com/articles/s44271-024-00114-4","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Surging interest in individual differences has faced setbacks in light of recent replication crises in psychology, for example in brain-wide association studies exploring brain-behavior correlations. A crucial component of replicability for individual differences studies, which is often assumed but not directly tested, is the reliability of the measures we use. Here, we evaluate the reliability of different cognitive tasks on a dataset with over 250 participants, who each completed a multi-day task battery. We show how reliability improves as a function of number of trials, and describe the convergence of the reliability curves for the different tasks, allowing us to score tasks according to their suitability for studies of individual differences. We further show the effect on reliability of measuring over multiple time points, with tasks assessing different cognitive domains being differentially affected. Data collected over more than one session may be required to achieve trait-like stability. Reliability of cognitive task measures improves as a function of number of trials. Because of differences in reliability convergence, tasks differ in suitability as estimates of individual differences. To achieve traitlike stability in measures, data must be combined across sessions.