Ömer Emre Can Alagöz, Thorsten Meiser, Lale Khorramdel
{"title":"Disentangling individual differences in cognitive response mechanisms for rating scale items: A flexible-mixture multidimensional IRTree approach.","authors":"Ömer Emre Can Alagöz, Thorsten Meiser, Lale Khorramdel","doi":"10.3758/s13428-025-02778-0","DOIUrl":"10.3758/s13428-025-02778-0","url":null,"abstract":"<p><p>The accuracy of our inferences from rating-scale items can be improved with IRTree models, which consider heuristic response strategies like response styles (RS). IRTree models break down ordinal responses into pseudo-items (nodes), each representing a distinct decision-making process. These nodes are then modeled using an item response model. In the case of four-point items, a response is split into two nodes: 1) response direction, where the trait influences the overall agreement with items, and 2) response extremity, where both the trait and extreme RS (ERS) impact the choice of relative (dis)agreement categories. However, traditional models, despite addressing RS effects, assume that all respondents follow an identical response strategy, where the selection of relative (dis)agreement categories is influenced by the trait and ERS to the same degree for all respondents. Given that respondents may vary in the extent to which they adopt heuristic-driven strategies (e.g., fatigue, motivation, expertise), this assumption of homogeneous response processes is unlikely to be satisfied, potentially leading to inaccurate inferences. To accommodate different response strategies, we introduce the mixture IRTree model (MixTree). In MixTree, participants are assigned to different latent classes, each associated with distinct response processes. Based on their class memberships, varying weights are assigned to individuals' trait and ERS scores. Additionally, MixTree simultaneously examines extraneous variables to explore sources of heterogeneity. A simulation study validates the MixTree's performance in recovering classes and model parameters. Empirical data analysis identifies two latent classes, one linked to a trait-driven and the other to RS-driven mechanisms.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 9","pages":"256"},"PeriodicalIF":3.9,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12343705/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144833870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A compendium of considerations for methods in motion-induced blindness research.","authors":"Vishnu Y Soni, Joey Planchet, John E Sparrow","doi":"10.3758/s13428-025-02789-x","DOIUrl":"10.3758/s13428-025-02789-x","url":null,"abstract":"<p><p>Ever since its introduction to vision research by Bonneh and colleagues in 2001, motion-induced blindness (MIB) has spawned a great deal of scholarly activity. However, an investigation of the common methods used for the MIB task indicates several issues, including a trend of newer studies simply replicating methodologies from earlier works, ambiguity regarding what the most optimal methods are for the MIB task, and the exclusion by MIB studies of crucial details regarding the MIB task. These issues are the consequence of the past two decades of MIB research having transpired without an updated set of guidelines and considerations regarding the use of the MIB task. Therefore, we aim to review the prevalent research methods in MIB studies, shed light on overlooked implications associated with these methodologies and configurations, and highlight innovative and effective approaches to MIB-related data collection. Discussions are supported using insights from both past MIB studies and new empirical findings of the present study. Aspects of MIB research that have been covered include managing individual differences, participant fatigue, fixation, response collection methods, defining target disappearance, target positioning, and the appearance of a MIB display's constituent elements.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 9","pages":"259"},"PeriodicalIF":3.9,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144844291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Boundaries in the eyes: Measure event segmentation during naturalistic video watching using eye tracking.","authors":"Jiashen Li, Zhengyue Chen, Xin Hao, Wei Liu","doi":"10.3758/s13428-025-02790-4","DOIUrl":"10.3758/s13428-025-02790-4","url":null,"abstract":"<p><p>During naturalistic information processing, individuals spontaneously segment their continuous experiences into discrete events, a phenomenon known as event segmentation. Traditional methods for assessing this process, which include subjective reports and neuroimaging techniques, often disrupt real-time segmentation or are costly and time-intensive. Our study investigated the potential of measuring event segmentation by recording and analyzing eye movements while participants viewed naturalistic videos. We collected eye movement data from healthy young adults as they watched commercial films (N = 104), or online Science, Technology, Engineering, and Mathematics (STEM) educational courses (N = 44). We analyzed changes in pupil size and eye movement speed near event boundaries and employed inter-subject correlation analysis (ISC) and hidden Markov models (HMM) to identify patterns indicative of event segmentation. We observed that both the speed of eye movements and pupil size dynamically responded to event boundaries, exhibiting heightened sensitivity to high-strength boundaries. Our analyses further revealed that event boundaries synchronized eye movements across participants. These boundaries can be effectively identified by HMM, yielding higher within-event similarity values and aligned with human-annotated boundaries. Importantly, HMM-based event segmentation metrics responded to experimental manipulations and predicted learning outcomes. This study provided a comprehensive computational framework for measuring event segmentation using eye-tracking. With the widespread accessibility of low-cost eye-tracking devices, the ability to measure event segmentation from eye movement data promises to deepen our understanding of this process in diverse real-world settings.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 9","pages":"255"},"PeriodicalIF":3.9,"publicationDate":"2025-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144833869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Peter D Kvam, Konstantina Sokratous, Anderson K Fitch, Jasmin Vassileva
{"title":"Comparing likelihood-based and likelihood-free approaches to fitting and comparing models of intertemporal choice.","authors":"Peter D Kvam, Konstantina Sokratous, Anderson K Fitch, Jasmin Vassileva","doi":"10.3758/s13428-025-02779-z","DOIUrl":"10.3758/s13428-025-02779-z","url":null,"abstract":"<p><p>Machine learning methods have recently begun to be used for fitting and comparing cognitive models, yet they have mainly focused on methods for dealing with models that lack tractable likelihoods. Evaluating how these approaches compare to traditional likelihood-based methods is critical to understanding the utility of machine learning for modeling and determining what role it might play in the development of new models and theories. In this paper, we systematically benchmark neural network approaches against likelihood-based approaches to model fitting and comparison, focusing on intertemporal choice modeling as an illustrative application. By applying each approach to intertemporal choice data from participants with substance use problems, we show that there is convergence between neural network and Bayesian methods when it comes to making inferences about latent processes and related substance use outcomes. For model comparison, however, classification networks significantly outperformed likelihood-based metrics. Next, we explored two extensions of this approach, using recurrent layers to allow them to fit data with variable stimuli and numbers of trials, and using dropout layers to allow for posterior sampling. We ultimately suggest that neural networks are better suited to fast parameter estimation and posterior sampling, applications to large data sets, and model comparison, while Bayesian MCMC methods should be preferred for flexible applications to smaller data sets featuring many conditions or experimental designs.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 9","pages":"252"},"PeriodicalIF":3.9,"publicationDate":"2025-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12339600/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144820453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"The polish vocabulary size test: A novel adaptive test for receptive vocabulary assessment.","authors":"Danil Fokin, Monika Płużyczka, Grigory Golovin","doi":"10.3758/s13428-025-02775-3","DOIUrl":"10.3758/s13428-025-02775-3","url":null,"abstract":"<p><p>We present the Polish Vocabulary Size Test (PVST), a novel tool for assessing the receptive vocabulary size of both native and non-native Polish speakers. Based on item response theory and computerized adaptive testing, PVST dynamically adjusts to each test-taker's proficiency level, ensuring high accuracy while keeping the test duration short. To validate the test, a pilot study was conducted with 1475 participants. Native Polish speakers demonstrated significantly larger vocabularies (mean = 75,125 words; range = 19,556-122,693) compared to non-native speakers (mean = 7165 words; range = 646-23,394). For native speakers, vocabulary size showed a strong positive correlation with age (r = .496, p < .001). The PVST is available online at myvocab.info/pl .</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 9","pages":"254"},"PeriodicalIF":3.9,"publicationDate":"2025-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12339579/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144820454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Cognitive modeling of lure discriminability in the Mnemonic Similarity Task.","authors":"Tianye Ma, Weiwei Zhang","doi":"10.3758/s13428-025-02785-1","DOIUrl":"10.3758/s13428-025-02785-1","url":null,"abstract":"<p><p>Lure discrimination in the Mnemonic Similarity Task (MST) has been widely used to measure pattern separation. However, the classic index of lure discrimination in the MST has arbitrary assumptions with limited supporting evidence. The present study has thus developed several models with different assumptions on the process underlying MST as well as the different model-derived indices of lure discrimination. Furthermore, we have assessed and compared these models in a measurement-based approach. We found that the model for the classic lure discrimination index fails to accurately predict the responses in an MST from > 150 participants. Instead, a new index based on the unidimensional signal detection model provides the best fits of the empirical dataset. This work highlights the value of model-based approaches in measuring lure discriminability in the MST.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 9","pages":"253"},"PeriodicalIF":3.9,"publicationDate":"2025-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144820452","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yang Ba, Michelle V Mancenido, Erin K Chiou, Rong Pan
{"title":"Data quality in crowdsourcing and spamming behavior detection.","authors":"Yang Ba, Michelle V Mancenido, Erin K Chiou, Rong Pan","doi":"10.3758/s13428-025-02757-5","DOIUrl":"10.3758/s13428-025-02757-5","url":null,"abstract":"<p><p>As crowdsourcing emerges as an efficient and cost-effective method for obtaining labels for machine learning datasets, it is important to assess the quality of crowd-provided data to improve analysis performance and reduce biases in subsequent machine learning tasks. Given the lack of ground truth in most cases of crowdsourcing, we refer to data quality as the annotators' consistency and credibility. Unlike the simple scenarios where kappa coefficient and intraclass correlation coefficient usually can apply, online crowdsourcing requires dealing with more complex situations. We introduce a systematic method for evaluating data quality and detecting spamming threats via variance decomposition, and we classify spammers into three categories based on their different behavioral patterns. A spammer index is proposed to assess entire data consistency, and two metrics are developed to measure crowd workers' credibility by utilizing the Markov chain and generalized random effects models. Furthermore, we demonstrate the practicality of our techniques and their advantages by applying them to a face verification task using both simulated and real-world data collected from two crowdsourcing platforms.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 9","pages":"251"},"PeriodicalIF":3.9,"publicationDate":"2025-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144798051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Testing measurement and structural invariance in latent mediation models - A comparison of IPCR and Bayesian MNLFA.","authors":"Fabian Felix Muench, Tobias Koch","doi":"10.3758/s13428-025-02781-5","DOIUrl":"10.3758/s13428-025-02781-5","url":null,"abstract":"<p><p>Moderated mediation models are frequently used in psychological research to examine direct, indirect, and total effects across an external moderating variable. When these models involve latent variables, measurement invariance should be tested first to ensure that measures function equivalently across subpopulations. If measurement invariance is violated, conclusions drawn about the moderation effects can be biased. However, measurement invariance is seldom tested across the moderator variable itself, especially if it is continuous. In this paper, we present two approaches that allow testing measurement and structural invariance simultaneously and across continuous covariates. They are termed individual parameter contribution regression (IPCR; Arnold et al., Structural Equation Modeling: A Multidisciplinary Journal, 27, 613-628, 2019) and moderated nonlinear latent factor analysis (MNLFA; Bauer & Hussong, Psychological Methods, 14(2), 101-125, 2009). We showcase both approaches with empirical data of <math><mrow><mi>N</mi> <mo>=</mo> <mn>399</mn></mrow> </math> couples in the German Family Panel (Brüderl et al., 2022). We show how MNLFA can be estimated in a Bayesian framework and explain Bayesian model selection with posterior predictive model checks and leave-one-out cross-validation (Vehtari et al., Statistics and Computing, 27(5), 1413-1432, 2017). Afterwards, we present the results of a simulation study comparing IPCR and Bayesian MNLFA with regard to parameter bias. We close with a comparison of both approaches regarding the empirical analysis and the simulation study and provide recommendations for applied researchers working with latent moderated mediation models.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 9","pages":"250"},"PeriodicalIF":3.9,"publicationDate":"2025-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12334378/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144798052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Aligning syntactic structure to the dynamics of verbal communication: A pipeline for annotating syntactic phrases onto speech acoustics.","authors":"Cosimo Iaia, Alessandro Tavano","doi":"10.3758/s13428-025-02747-7","DOIUrl":"10.3758/s13428-025-02747-7","url":null,"abstract":"<p><p>To investigate how the human brain encodes the complex dynamics of natural languages, any viable and reproducible analysis pipeline must rely on either manual annotations or natural language processing (NLP) tools, which extract relevant physical (e.g., acoustic, gestural), and structure-building information from speech and language signals. However, annotating syntactic structure for a given natural language is arguably a harder task than annotating the onset and offset of speech units such as phonemes and syllables, as the latter can be identified by relying on the physically overt and temporally measurable properties of the signal, while syntactic units are generally covert and their chunking is model-driven. We describe and validate a pipeline that takes into account both physical and theoretical aspects of speech and language signals, and operates a theory-driven and explicit alignment between overt speech units and covert syntactic units.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 9","pages":"249"},"PeriodicalIF":3.9,"publicationDate":"2025-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12331764/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144798050","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Merlin Monzel, Christian O Scholz, Joel Pearson, Martin Reuter
{"title":"Why indecisive trials matter: Improving the binocular rivalry imagery priming score for the assessment of aphantasia.","authors":"Merlin Monzel, Christian O Scholz, Joel Pearson, Martin Reuter","doi":"10.3758/s13428-025-02780-6","DOIUrl":"10.3758/s13428-025-02780-6","url":null,"abstract":"<p><p>Since mental imagery cannot be observed from the outside, it is all the more important to make it measurable. Yet, many so-called mental imagery tasks confuse object and spatial imagery or can be solved entirely without mental imagery, making them inappropriate for the assessment of mental imagery strength. One promising measurement method is the binocular rivalry task by Pearson et al. (Current Biology 18(13):982-986, 2008), which uses mental imagery priming to quantify mental imagery strength. Here, we propose an improved equation for the binocular rivalry priming score to significantly increase its predictive validity. In a sample of 38 aphantasics and 73 controls, we demonstrate that the binocular rivalry priming score calculated by the new equation explains more variance in the self-reported mental imagery capacity than the original equation. The improved binocular rivalry priming score is particularly beneficial when only a few trials are recorded (e.g., due to time or attention constraints) or when people with low mental imagery (i.e., aphantasics) have to be identified. The improved binocular rivalry priming score is advantageous in many situations, making it the preferred measure for future research.</p>","PeriodicalId":8717,"journal":{"name":"Behavior Research Methods","volume":"57 9","pages":"248"},"PeriodicalIF":3.9,"publicationDate":"2025-08-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12321684/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144783369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}