AssessmentPub Date : 2025-03-01Epub Date: 2025-02-10DOI: 10.1177/10731911251320053
Kristian E Markon, Thomas Olino
{"title":"From Confirmation to Theory: Introduction to the Special Section on \"Assessment, Measurement, and Registered Replication\".","authors":"Kristian E Markon, Thomas Olino","doi":"10.1177/10731911251320053","DOIUrl":"10.1177/10731911251320053","url":null,"abstract":"<p><p>The goal of this special section was to encourage discussion about replication in assessment science, as well as the use and development of registered replication and open science more broadly. The contributions span a range of topics, from the replicability and generalizability of measurement models, to challenges in the replicability of results from intensively longitudinal designs, to a discussion of how preregistration should be used in assessment science. Collectively, the papers in this special section provide a view of replication as part of the process by which scientific theories are evaluated and evolve.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"167-169"},"PeriodicalIF":3.5,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143389702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
AssessmentPub Date : 2025-03-01Epub Date: 2024-05-23DOI: 10.1177/10731911241253430
E David Klonsky
{"title":"Campbell's Law Explains the Replication Crisis: Pre-Registration Badges Are History Repeating.","authors":"E David Klonsky","doi":"10.1177/10731911241253430","DOIUrl":"10.1177/10731911241253430","url":null,"abstract":"<p><p>Campbell's Law explains the replication crisis. In brief, useful tools such as hypotheses, <i>p</i>-values, and multi-study designs came to be viewed as indicators of strong science, and thus goals in and of themselves. Consequently, their use became distorted in unanticipated ways (e.g., hypothesizing after results were known [HARKing], p-Hacking, misuses of researcher degrees of freedom), and fragile findings proliferated. Pre-registration mandates are positioned as an antidote. However, I argue that such efforts, perhaps best exemplified by pre-registration badges (PRBs), are history repeating: Another useful tool has been converted into an indicator of strong science and a goal in and of itself. This, too, will distort its use and harm psychological science in unanticipated ways. For example, there is already evidence that papers seeking PRBs routinely violate the rules and spirit of pre-registration. I suggest that pre-registration mandates will (a) discourage optimal scientific practice, (b) exacerbate the file drawer problem, (c) encourage pre-registering after results are known (PRARKing), and (d) create false trust in fragile findings. I conclude that multiple design features can help support replicability (e.g., adequate sample size, valid measurement, robustness checks, pre-registration), none should be canonized, replication is the only arbiter of replicability, and the most important solution is sociocultural: to foster a field that reveres and reinforces robust science-just as we once revered and reinforced flashy but fragile science.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"224-234"},"PeriodicalIF":3.5,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11874590/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141086411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
AssessmentPub Date : 2025-03-01Epub Date: 2024-11-14DOI: 10.1177/10731911241286868
Ginette Lafit, Jordan Revol, Leonie Cloos, Peter Kuppens, Eva Ceulemans
{"title":"The Effect of Different Construct Operationalizations, Study Duration, and Preprocessing Choices on Power-Based Sample Size Recommendations in Intensive Longitudinal Research.","authors":"Ginette Lafit, Jordan Revol, Leonie Cloos, Peter Kuppens, Eva Ceulemans","doi":"10.1177/10731911241286868","DOIUrl":"10.1177/10731911241286868","url":null,"abstract":"<p><p>To shed light on the dynamics of psychological processes, researchers often collect intensive longitudinal (IL) data by asking people to repeatedly report on their momentary experiences in daily life. Two important decisions when designing an IL study concern the number of persons and the number of measurement occasions to be included. These sample size decisions are ideally based on statistical power considerations. When conducting statistical power analysis, it is necessary to provide the value of the effect size of interest as well as of all other model parameters. In IL research, these values have to be based on previous studies. This implies that these values are subject to large heterogeneity due to differences in study design and preprocessing choices. This between-study heterogeneity can severely impact power-based sample size recommendations. In this article, we introduce an approach to investigate the impact of study design and pre-processing of previous studies and to determine a recommended sample size to account for this impact. We demonstrate how to use this approach to investigate the effect of different construct operationalizations, study duration, and preprocessing choices. This approach paves the way for more thoughtful and robust sample-size decisions.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"206-223"},"PeriodicalIF":3.5,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142613862","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
AssessmentPub Date : 2025-03-01Epub Date: 2024-12-14DOI: 10.1177/10731911241301475
Ashmita Ghosh, Nathaniel L Phillips, Chelsea E Sleep, Donald R Lynam, Joshua D Miller
{"title":"Examining the Structure of Personality Dysfunction Among Individuals Involved in Mental Health treatment: A Registered Report.","authors":"Ashmita Ghosh, Nathaniel L Phillips, Chelsea E Sleep, Donald R Lynam, Joshua D Miller","doi":"10.1177/10731911241301475","DOIUrl":"10.1177/10731911241301475","url":null,"abstract":"<p><p>Personality dysfunction is a core element of the diagnosis of personality disorders in both main diagnostic systems (<i>Diagnostic and Statistical Manual of Mental Disorders</i> [5th ed.] Personality Disorders and <i>ICD-11</i> [<i>International Classification of Diseases</i>, 11th Revision] Personality Disorders). A recent study by Sleep et al. explored the structure of personality dysfunction (<i>N</i> = 517 undergraduates) with exploratory bass-ackward factor analyses at the item level using items from six measures of personality functioning. The present study represents a replication of the study by Sleep and colleagues in a community sample of individuals currently in or seeking psychological/psychiatric treatment. Participants were recruited from Prolific who completed items from six measures of personality dysfunction (<i>N</i> = 457). The primary factor analysis identified three lower-order factors of impairment (Negative Self-Regard, Disagreeableness, and Intimacy Problems). These dysfunction factors showed moderate to strong relations to the factors generated by Sleep et al.; however, there were important differences in the overall structure suggesting a lack of formal replication.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"283-299"},"PeriodicalIF":3.5,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142823622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
AssessmentPub Date : 2025-02-28DOI: 10.1177/10731911251316468
Alexandra Hernandez-Vallant, Margo C Hurlocker, Alison Looby, Byron L Zamboanga
{"title":"The Structural Validity of the Barkley Deficits in Executive Functioning Scale-Short Form Among College Students Who Drink Alcohol.","authors":"Alexandra Hernandez-Vallant, Margo C Hurlocker, Alison Looby, Byron L Zamboanga","doi":"10.1177/10731911251316468","DOIUrl":"https://doi.org/10.1177/10731911251316468","url":null,"abstract":"<p><p>The Barkley Deficits in Executive Functioning Scale (BDEFS) is a multidimensional measure of executive functioning (EF). Yet, no work has validated the 20-item BDEFS short form (BDEFS-SF) among individuals at risk of EF deficits, including college students who consume alcohol. The current study evaluated the factor structure and measurement invariance of the BDEFS-SF in 867 undergraduates from 12 colleges/universities who endorsed past-month alcohol use. The bifactor model best fit the data, supporting the use of the general factor over the five subscales for measurement models, and measurement invariance was partially supported across sex assigned at birth. Four subscales were uniquely associated with alcohol use, consequences, and alcohol protective behavioral strategies. The BDEFS-SF total score appears to be a valid measure of EF among undergraduates who use alcohol; however, more research is needed to increase the clinical utility of the measure given the lack of support for the use of the subscales in measurement models.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251316468"},"PeriodicalIF":3.5,"publicationDate":"2025-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143530971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
AssessmentPub Date : 2025-02-24DOI: 10.1177/10731911251317785
G Leonard Burns, Juan José Montaño, Stephen P Becker, Mateu Servera
{"title":"Psychometric and Normative Information on the Child and Adolescent Behavior Inventory With Parent Ratings in a Nationally Representative Sample of Spanish Youth.","authors":"G Leonard Burns, Juan José Montaño, Stephen P Becker, Mateu Servera","doi":"10.1177/10731911251317785","DOIUrl":"https://doi.org/10.1177/10731911251317785","url":null,"abstract":"<p><p>Psychometric and normative information is provided for the Child and Adolescent Behavior Inventory (CABI) cognitive disengagement syndrome, anxiety, depression, attention-deficit/hyperactivity disorder (ADHD)-inattention, ADHD-hyperactivity/impulsivity, oppositional defiant disorder, social impairment, peer rejection, withdrawal from peer interactions, and academic impairment scales with a nationally representative sample of Spanish youth. Parents of 5,525 Spanish youth (ages 5-16, 56.1% males) completed the CABI scales on their sons and daughters. Scores on the 10 CABI scales demonstrated excellent reliability, invariance, and validity for males and females within early childhood (ages 5-8), middle childhood (ages 9-12), and adolescence (ages 13-16). Normative information (<i>T</i>-scores) is provided for females and males within each age group for the 10 CABI scales. The new psychometric and normative information increase the usefulness of the CABI scale scores for research and clinical activities. Copies of the CABI and the norms are available at no cost to professionals.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251317785"},"PeriodicalIF":3.5,"publicationDate":"2025-02-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143481990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
AssessmentPub Date : 2025-02-19DOI: 10.1177/10731911251317788
Cory A Munroe, Jennifer Leckey, Shannon A Johnson, Sophie Jacques
{"title":"Factor Structure and Psychometric Properties of the Learning, Executive, and Attention Functioning (LEAF) Scale in Young Adults.","authors":"Cory A Munroe, Jennifer Leckey, Shannon A Johnson, Sophie Jacques","doi":"10.1177/10731911251317788","DOIUrl":"https://doi.org/10.1177/10731911251317788","url":null,"abstract":"<p><p>The Learning, Executive, and Attention Functioning (LEAF) scale is a resource-friendly means of assessing executive functions (EFs) and related constructs (e.g., academic abilities) in children and adolescents that has been adapted for use with adults. However, no study in any population has investigated the factor structure of all LEAF EF items to determine whether items factor in a manner consistent with the originally proposed scale structure. Therefore, we examined LEAF scale responses of 546 young adults (<i>M</i><sub>age</sub> = 20.05, <i>SD</i> = 2.17). Upon removing academic items following a preliminary factor analysis, we performed principal axis factoring on the remaining 39 EF items. The final model accounted for 61.75% of the total variance in LEAF EF items and suggested that these items assess six moderately correlated EF constructs in young adults. We constructed six updated subscales to help researchers measure these EFs in young adults using the LEAF scale, each of which uniquely and differentially predicted measures of self-reported impulsivity, academic difficulties, and learning-related disorder history. Overall, the LEAF promises to be an accessible means of assessing a range of EF constructs in young adults, particularly when updated subscale structures based on factor analysis are used.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251317788"},"PeriodicalIF":3.5,"publicationDate":"2025-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143456728","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Assessment of Cognition and Language Using Alternative Response Modalities.","authors":"Kristine Stadskleiv, Katy Latham, Kristina Tufteskog Spanne, Karen Sætre, Anna Fraas, Ilaria Ruscito, Yasmine Taha, Janice Murray","doi":"10.1177/10731911251315012","DOIUrl":"https://doi.org/10.1177/10731911251315012","url":null,"abstract":"<p><p>Assessing cognition and language using standardized tests is challenging when the individual has severe speech and motor impairments. Tests with a multiple-choice format may be adapted without compromising standardization using alternative response modes like partner assisted scanning (PAS) and eye-pointing (EP). Standardization of such assessment is little researched. The study investigates the (a) reliability of, (b) transparency of, and (c) adherence to assessment protocols using PAS and EP. The participants were students from special needs education and speech and language therapy, who worked in dyads (<i>n</i> = 39). Two observers recorded a number of errors made in protocol delivery, independently of each other. The dyads made between 0 and 81.5 errors. Number of errors was not related to response mode, <i>t</i>(38) = -0.21, <i>p</i> = .839. The observers were in high agreement, with an intraclass correlation coefficient of .97, <i>p</i> < .001. The study suggests that assessing language involving alternative modes of responding can be successfully taught to novice practitioners.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251315012"},"PeriodicalIF":3.5,"publicationDate":"2025-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143397928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
AssessmentPub Date : 2025-02-10DOI: 10.1177/10731911251315626
Michael D Barnett, Jenna M Moore, Rebekah A Griffin
{"title":"Development of an Examiner-Examinee Rapport Scale for Neuropsychological Testing: The Barnett Rapport Questionnaire - 2.","authors":"Michael D Barnett, Jenna M Moore, Rebekah A Griffin","doi":"10.1177/10731911251315626","DOIUrl":"https://doi.org/10.1177/10731911251315626","url":null,"abstract":"<p><p>Rapport in the context of neuropsychological assessment refers to the interpersonal interaction between the examinee and examiner. The purpose of this study was to improve upon the original Barnett Rapport Questionnaire. Participants consisted of college students (<i>n</i> = 55), middle-aged adults (<i>n</i> = 30), and older adults (<i>n</i> = 99) at a university-affiliated neuropsychology clinic, who were administered a brief test battery. After testing, both examinees (<i>n</i> = 184) and examiners (<i>n</i> = 12) completed items regarding the nature of the interpersonal interaction. Confirmatory factor analyses indicated a two-factor model consisting of nine items found on both the examinee and examiner versions. No age cohort differences were found on the total Barnett Rapport Questionnaire - 2 (BRQ-2) scores for the examiner or examinee version; however, older adults endorsed the positively worded items more than young adults, and examiners endorsed the negatively worded items more for young adult examinees than for older adult examinees. No differences were found by gender or by race/ethnicity. The BRQ-2 promises better usability and improved psychometric qualities for the measurement of rapport in the context of neuropsychological assessment.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251315626"},"PeriodicalIF":3.5,"publicationDate":"2025-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143381444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
AssessmentPub Date : 2025-02-03DOI: 10.1177/10731911251315027
Salome D Odermatt, Silvia Grieder, Florine Schweizer, Anette Bünger, Alexander Grob
{"title":"The Role of Language Aspects in the Assessment of Cognitive and Developmental Functions in Children: An Analysis of the Intelligence and Development Scales-2.","authors":"Salome D Odermatt, Silvia Grieder, Florine Schweizer, Anette Bünger, Alexander Grob","doi":"10.1177/10731911251315027","DOIUrl":"https://doi.org/10.1177/10731911251315027","url":null,"abstract":"<p><p>The assessment of multilingual participants is challenging, as, for example, proficiency in the test language may interfere with test performance. We examined whether different language aspects (i.e., receptive and expressive language abilities, bi/multilingualism) contribute independently to the variance in scores on cognitive and developmental functions of the Intelligence and Development Scales-2 (IDS-2). The sample comprised 826 children aged 5 to 10 years living in German-speaking regions. Hierarchical regression analyses revealed that receptive language ability was significantly associated with almost all IDS-2 scores. Expressive language ability explained little additional variance, except for the intelligence composites, Verbal Reasoning (including subtests), and the basic skills subtests. Bi/multilingualism explained variance above language abilities only in subtests of Verbal Reasoning and verbal Long-Term Memory. Findings highlight the importance of considering language proficiency, particularly in tasks with high verbal demands, when assessing cognitive and developmental functions with the IDS-2 in participants at risk for linguistic disadvantages.</p>","PeriodicalId":8577,"journal":{"name":"Assessment","volume":" ","pages":"10731911251315027"},"PeriodicalIF":3.5,"publicationDate":"2025-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143078502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}