Reza de Souza Brümmer, Karolin Rose Krause, Giovanni Abrahão Salum, Marcelo Pio de Almeida Fleck, Ighor Miron Porto, João Villanova do Amaral, João Pedro Gonçalves Pacheco, Bettina Moltrecht, Eoin McElroy, Mauricio Scopel Hoffmann
{"title":"Beyond frequency: Evaluating the validity of assessing the context, duration, ability, and botherment of depression and anxiety symptoms in South Brazil.","authors":"Reza de Souza Brümmer, Karolin Rose Krause, Giovanni Abrahão Salum, Marcelo Pio de Almeida Fleck, Ighor Miron Porto, João Villanova do Amaral, João Pedro Gonçalves Pacheco, Bettina Moltrecht, Eoin McElroy, Mauricio Scopel Hoffmann","doi":"10.1037/pas0001323","DOIUrl":"10.1037/pas0001323","url":null,"abstract":"<p><p>Assessment tools for depression and anxiety usually inquire about the frequency of symptoms. However, evidence suggests that different question framings might trigger different responses. Our aim is to test if asking about symptom's context, ability, duration, and botherment adds validity to Patient Health Questionnaire-9, General Anxiety Disorder-7, and Patient-Related Outcome Measurement Information Systems depression and anxiety. Participants came from two cross-sectional convenience-sampled surveys (<i>N</i> = 1,871) of adults (66% females, aged 33.4 ± 13.2), weighted to approximate with the state-level population. We examined measurement invariance across the different question frames, estimated whether framing affected mean scores, and tested their independent validity using covariate-adjusted and sample-weighted structural equation models. Validity was tested using tools assessing general disability, alcohol use, loneliness, well-being, grit, and frequency-based questions from depression and anxiety questionnaires. A bifactor model was applied to test the internal consistency of the question frames under the presence of a general factor (i.e., depression or anxiety). Measurement invariance was supported across the different frames. Framing questions as ability (i.e., \"How easily …\") produced a higher score, compared with framing by context (i.e., \"In which daily situations …\"). Construct and criterion validity analysis demonstrate that variance explained using multiple question frames was similar to using only one. We detected a strong overarching factor for each instrument, with little variances left to be explained by the question frame. Therefore, it is unlikely that using different adverbial phrasings can help clinicians and researchers to improve their ability to detect depression or anxiety. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"488-504"},"PeriodicalIF":3.3,"publicationDate":"2024-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141458963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Estimating classification consistency of machine learning models for screening measures.","authors":"Oscar Gonzalez, A R Georgeson, William E Pelham","doi":"10.1037/pas0001313","DOIUrl":"10.1037/pas0001313","url":null,"abstract":"<p><p>This article illustrates novel quantitative methods to estimate classification consistency in machine learning models used for screening measures. Screening measures are used in psychology and medicine to classify individuals into diagnostic classifications. In addition to achieving high accuracy, it is ideal for the screening process to have high classification consistency, which means that respondents would be classified into the same group every time if the assessment was repeated. Although machine learning models are increasingly being used to predict a screening classification based on individual item responses, methods to describe the classification consistency of machine learning models have not yet been developed. This article addresses this gap by describing methods to estimate classification inconsistency in machine learning models arising from two different sources: sampling error during model fitting and measurement error in the item responses. These methods use data resampling techniques such as the bootstrap and Monte Carlo sampling. These methods are illustrated using three empirical examples predicting a health condition/diagnosis from item responses. R code is provided to facilitate the implementation of the methods. This article highlights the importance of considering classification consistency alongside accuracy when studying screening measures and provides the tools and guidance necessary for applied researchers to obtain classification consistency indices in their machine learning research on diagnostic assessments. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":"36 6-7","pages":"395-406"},"PeriodicalIF":3.3,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11952017/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141200498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hannah N Wyant, Marc A Silva, Stephanie Agtarap, Farina A Klocksieben, Teagen Smith, Risa Nakase-Richardson, Shannon R Miles
{"title":"Applying the PTSD Checklist-Civilian and PTSD Checklist for DSM-5 crosswalk in a traumatic brain injury sample: A veterans affairs traumatic brain injury model systems study.","authors":"Hannah N Wyant, Marc A Silva, Stephanie Agtarap, Farina A Klocksieben, Teagen Smith, Risa Nakase-Richardson, Shannon R Miles","doi":"10.1037/pas0001315","DOIUrl":"10.1037/pas0001315","url":null,"abstract":"<p><p>This study evaluates the use of the crosswalk between the PTSD Checklist-Civilian (PCL-C) and PTSD Checklist for DSM-5 (PCL-5) designed by Moshier et al. (2019) in a sample of service members and veterans (SM/V; N = 298) who had sustained a traumatic brain injury (TBI) and were receiving inpatient rehabilitation. The PCL-C and PCL-5 were completed at the same time. Predicted PCL-5 scores for the sample were obtained according to the crosswalk developed by Moshier et al. We used three measures of agreement: intraclass correlation coefficient (ICC), mean difference between predicted and observed scores, and Cohen's κ to determine the performance of the crosswalk in this sample. Subgroups relevant to those who have sustained a TBI, such as TBI severity, were also examined. There was strong agreement between the predicted and observed PCL-5 scores (ICC = .95). The overall mean difference between predicted and observed PCL-5 scores was 0.07 and not statistically significant (SD = 8.29, p = .89). Significant mean differences between predicted and observed PCL-5 scores calculated between subgroups were seen in Black participants (MD = -4.09, SD = 8.41, p = .01) and those in the Year 5 follow-up group (MD = 1.77, SD = 7.14, p = .03). Cohen's κ across subgroups had a mean of κ = 0.76 (.57-1.0), suggesting that there was moderate to almost perfect diagnostic agreement. Our results suggest the crosswalk created by Moshier et al. can be applied to SM/V who have suffered a TBI. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":"36 6-7","pages":"425-432"},"PeriodicalIF":3.3,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141200380","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Evelien Schat, Francis Tuerlinckx, Marieke J Schreuder, Bart De Ketelaere, Eva Ceulemans
{"title":"Forecasting the onset of depression with limited baseline data only: A comparison of a person-specific and a multilevel modeling based exponentially weighted moving average approach.","authors":"Evelien Schat, Francis Tuerlinckx, Marieke J Schreuder, Bart De Ketelaere, Eva Ceulemans","doi":"10.1037/pas0001314","DOIUrl":"10.1037/pas0001314","url":null,"abstract":"<p><p>The onset of depressive episodes is preceded by changes in mean levels of affective experiences, which can be detected using the exponentially weighted moving average procedure on experience sampling method (ESM) data. Applying the exponentially weighted moving average procedure requires sufficient baseline data from the person under study in healthy times, which is needed to calculate a control limit for monitoring incoming ESM data. It is, however, not trivial to obtain sufficient baseline data from a single person. We therefore investigate whether historical ESM data from healthy individuals can help establish an adequate control limit for the person under study via multilevel modeling. Specifically, we focus on the case in which there is very little baseline data available of the person under study (i.e., up to 7 days). This multilevel approach is compared with the traditional, person-specific approach, where estimates are obtained using the person's available baseline data. Predictive performance in terms of Matthews correlation coefficient did not differ much between the approaches; however, the multilevel approach was more sensitive at detecting mean changes. This implies that for low-cost and nonharmful interventions, the multilevel approach may prove particularly beneficial. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":"36 6-7","pages":"379-394"},"PeriodicalIF":3.3,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141200561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Examining the validity and factor structure of the ICD-11 trait domains.","authors":"Tiffany A Brown, Martin Sellbom","doi":"10.1037/pas0001308","DOIUrl":"10.1037/pas0001308","url":null,"abstract":"<p><p>The International Classification of Diseases, 11th Edition (ICD-11) includes a new personality disorder (PD) severity diagnosis that may be further characterized using up to five trait domain specifiers. Most of the previous studies have investigated the ICD-11 trait domains using self-report measures. The present study aimed to validate ICD-11 PD trait domains using a multimethod design in a community mental health sample (n = 336). We conducted two confirmatory factor analyses to examine the factor structure of the ICD-11 PD trait model, utilizing clinician-rating, self-report, and informant-report measures. Finally, we examined associations between clinician-rated, self-reported, and informant-reported ICD-11 trait domains with external criteria, specifically traditional PD symptoms and the five-factor model of normal personality. All clinician-rated, self-reported, and informant-reported domain scores loaded meaningfully on their expected factors when controlling for nontrivial method factors. Generally, the trait domains exhibited meaningful associations with conceptually relevant external criteria, although the anankastia domain exhibited more variability in its pattern of correlations across methods. Overall, the ICD-11 trait domain model shows promising reliability and validity, indicating good progress within the field of PD assessment toward a more useful PD operationalization. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":"36 5","pages":"311-322"},"PeriodicalIF":3.6,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140874789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Leon P Wendt, Johannes Zimmermann, Carsten Spitzer, Sascha Müller
{"title":"Mindreading measures misread? A multimethod investigation into the validity of self-report and task-based approaches.","authors":"Leon P Wendt, Johannes Zimmermann, Carsten Spitzer, Sascha Müller","doi":"10.1037/pas0001310","DOIUrl":"10.1037/pas0001310","url":null,"abstract":"<p><p>Mindreading ability-also referred to as cognitive empathy or mentalizing-is typically conceptualized as a relatively stable dimension of individual differences in the ability to make accurate inferences about the mental states of others. This construct is primarily assessed using self-report questionnaires and task-based performance measures. However, the validity of these measures has been questioned: According to rival interpretations, mindreading tasks may capture general cognitive ability, whereas mindreading self-reports may capture perceived rather than actual mindreading ability. In this preregistered multimethod study involving 700 participants from the U.S. general population, we tested the validity of mindreading measures by examining the nomological network of self-reports and task-based methods using structural equation modeling. Specifically, we contrasted the empirical associations with theoretical predictions that assume mindreading measures are valid versus invalid. More consistent with rival interpretations, mindreading tasks showed a negligible latent correlation with mindreading self-reports (.05) and a large one with general cognitive ability (.85), whereas mindreading self-reports were specifically associated with perceived performance in mindreading tasks (.29). Also more consistent with rival interpretations, neither mindreading self-reports nor task-based measures showed positive unique associations with psychosocial functioning when controlling for general cognitive ability and general positive self-evaluation. Instead, negative unique associations emerged for both methods, although this effect was not robust for tasks. Overall, the results cast doubt on the validity of commonly used mindreading measures and support their rival interpretations. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"365-378"},"PeriodicalIF":3.6,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139997236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Martin Sellbom, Tiffany A Brown, Mark H Waugh, Christopher J Hopwood
{"title":"Development and initial validation of Personality Disorder Syndrome scales for the MMPI-3.","authors":"Martin Sellbom, Tiffany A Brown, Mark H Waugh, Christopher J Hopwood","doi":"10.1037/pas0001311","DOIUrl":"10.1037/pas0001311","url":null,"abstract":"<p><p>The purpose of the present study was to revise and update the MMPI-2-RF personality disorder (PD) syndrome scales for the Minnesota Multiphasic Personality Inventory-3 (MMPI-3). Study 1 describes the development of the MMPI-3 PD syndrome scales in three separate samples of community participants (n = 1,591), university students (n = 1,660), and outpatient mental health patients (n = 1,537). The authors independently evaluated each of the 72 new MMPI-3 items and rated them for appropriateness for scale inclusion and used various statistical procedures for final item selection. Ultimately, all 10 scales were revised, with nine incorporating items that were new to the MMPI-3. In Study 2, we subsequently validated the new MMPI-3 PD Syndrome scales against measures of traditional PD measures, trait measures of the Alternative Diagnostic and Statistical Manual of Mental Disorders, fifth edition-5 model of personality disorders (AMPD) and the five-factor model (FFM) of personality, and specific criterion measures of externalizing, psychopathy, narcissism, emotional dysregulation, and self-harm, in two samples of university students (ns = 489 and 645). With some exceptions, the results were generally supportive of the convergent and discriminant validities of the MMPI-3 PD Syndrome scales. The Histrionic PD scale in particular was associated with questionable results and diverged most strongly from the theoretical construct it was originally meant to reflect. Further continuous validation of the scales is needed, especially in clinical samples, but the findings to date are promising. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":"36 5","pages":"323-338"},"PeriodicalIF":3.6,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140851187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Development and initial validation of an open-access online Behavioral Avoidance Test (BAT) for spider fear.","authors":"Markus Grill, Martin Heller, Anke Haberkamp","doi":"10.1037/pas0001305","DOIUrl":"https://doi.org/10.1037/pas0001305","url":null,"abstract":"<p><p>The behavioral avoidance test (BAT) is a well-known diagnostic tool assessing fear by directly measuring avoidance behavior. For instance, in spider phobia, participants or patients gradually approach a live spider until they feel too uncomfortable to continue. However, the use of different BAT protocols in various studies hampers the comparability of results. Moreover, conducting the test requires considerable preparation by researchers and clinicians. Thus, we have developed an open-access online BAT (vBATon). We validated its efficacy in measuring avoidance behavior and eliciting feelings of anxiety and disgust by comparing it to a real-life BAT (rl-BAT). Spider-fearful (N = 31) and nonfearful (N = 31) individuals completed a rl-BAT and vBATon on two separate dates within a 1-week interval. As expected, both tests successfully distinguished between spider-fearful and nonfearful individuals. Crucially, equivalence tests confirmed that vBATon captures avoidance behavior, anxiety, and disgust equal to the rl-BAT. Assessing validity, we found moderate to high correlations between vBATon and (a) the rl-BAT and (b) self-report measurements of spider fear (Spider Phobia Questionnaire, Fear of Spiders Questionnaire). Overall, our study displayed initial evidence of validity of vBATon and suggests that it is a standardized, efficient, and user-friendly alternative to rl-BATs for measuring spider fear. It can be utilized in both research and clinical practice. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":"36 5","pages":"351-364"},"PeriodicalIF":3.6,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140852721","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Investigating racial disparities in violence risk assessment using the Spousal Assault Risk Assessment Guide-Version 3 (SARA-V3): Structured professional judgment ratings and recidivism among Indigenous and non-Indigenous individuals.","authors":"Neil R Hogan, Gabriela Corăbian","doi":"10.1037/pas0001307","DOIUrl":"10.1037/pas0001307","url":null,"abstract":"<p><p>Racial disparities in criminal justice outcomes are widely observed. In Canada, such disparities are particularly evident between Indigenous and non-Indigenous persons. The role of formal risk assessment in contributing to such disparities remains a topic of interest to many, but critical analysis has almost exclusively focused on actuarial or statistical risk measures. Recent research suggests that ratings from other common tools, based on the structured professional judgment model, can also demonstrate racial disparities. This study examined risk assessments produced using a widely used structured professional judgment tool, the Spousal Assault Risk Assessment Guide-Version 3, among a sample of 190 individuals with histories of intimate partner violence. We examined the relationships among race, risk factors, summary risk ratings, and recidivism while also investigating whether participants' racial identity influenced the likelihood of incurring formal sanctions for reported violence. Spousal Assault Risk Assessment Guide-Version 3 risk factor totals and summary risk ratings were associated with new violent charges. Indigenous individuals were assessed as demonstrating more risk factors and were more likely to be rated as high risk, even after controlling for summed risk factor totals and prior convictions. They were also more likely to recidivate and to have a history of at least one reported act of violence that did not result in formal sanctions. The results suggest that structured professional judgment guidelines can produce disparate results across racial groups. The disparities observed may reflect genuine differences in the likelihood of recidivism, driven by psychologically meaningful risk factors which have origins in deep-rooted systemic and contextual factors. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"339-350"},"PeriodicalIF":3.6,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140176151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Min Eun Jeon, Nikhila S Udupa, Miracle R Potter, Morgan Robison, Lee Robertson, Megan L Rogers, Thomas E Joiner
{"title":"Measurement invariance of the Depressive Symptom Inventory-Suicidality Scale across race, ethnicity, sexual orientation, and plurality of minoritized identities.","authors":"Min Eun Jeon, Nikhila S Udupa, Miracle R Potter, Morgan Robison, Lee Robertson, Megan L Rogers, Thomas E Joiner","doi":"10.1037/pas0001306","DOIUrl":"10.1037/pas0001306","url":null,"abstract":"<p><p>Measurement invariance across diverse groups, though crucial for determining the generalizability of a measure, has not yet been tested in many assessments of suicidal thoughts. The present study assessed the measurement invariance and psychometric properties of one such assessment, the Depressive Symptom Inventory-Suicidality Subscale, across multiple identity dimensions in a large data set (<i>n</i> = 1,118) that combined three diverse samples. Findings supported measurement invariance of the scale by race, ethnicity, and sexual orientation when comparing those who were majority-aligning with their minoritized counterparts, as well as good internal consistency and expected convergent validity. The expected one-factor structure fit well for all three of the samples assessed. Overall, the results support measurement invariance and generalizability of the Depressive Symptom Inventory-Suicidality Subscale. Future studies should continue assessing measures of suicidal thoughts and behaviors through testing invariance across identities, especially as it pertains to specific identity subgroups and their intersections. (PsycInfo Database Record (c) 2024 APA, all rights reserved).</p>","PeriodicalId":20770,"journal":{"name":"Psychological Assessment","volume":" ","pages":"303-310"},"PeriodicalIF":3.3,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11696994/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139707684","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}