Xiaojian Sun, Tongxin Zhang, Chang Nie, Naiqing Song, Tao Xin
{"title":"Combining regularization and logistic regression model to validate the Q-matrix for cognitive diagnosis model","authors":"Xiaojian Sun, Tongxin Zhang, Chang Nie, Naiqing Song, Tao Xin","doi":"10.1111/bmsp.12346","DOIUrl":"10.1111/bmsp.12346","url":null,"abstract":"<p>Q-matrix is an important component of most cognitive diagnosis models (CDMs); however, it mainly relies on subject matter experts' judgements in empirical studies, which introduces the possibility of misspecified q-entries. To address this, statistical Q-matrix validation methods have been proposed to aid experts' judgement. A few of these methods, including the multiple logistic regression-based (MLR-B) method and the Hull method, can be applied to general CDMs, but they are either time-consuming or lack accuracy under certain conditions. In this study, we combine the <i>L</i><sub>1</sub> regularization and MLR model to validate the Q-matrix. Specifically, an <i>L</i><sub>1</sub> penalty term is imposed on the log-likelihood of the MLR model to select the necessary attributes for each item. A simulation study with various factors was conducted to examine the performance of the new method against the two existing methods. The results show that the regularized MLR-B method (a) produces the highest Q-matrix recovery rate (QRR) and true positive rate (TPR) for most conditions, especially with a small sample size; (b) yields a slightly higher true negative rate (TNR) than either the MLR-B or the Hull method for most conditions; and (c) requires less computation time than the MLR-B method and similar computation time as the Hull method. A real data set is analysed for illustration purposes.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"78 1","pages":"1-21"},"PeriodicalIF":1.5,"publicationDate":"2024-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140677462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Three new corrections for standardized person-fit statistics for tests with polytomous items","authors":"Kylie Gorney","doi":"10.1111/bmsp.12342","DOIUrl":"10.1111/bmsp.12342","url":null,"abstract":"<p>Recent years have seen a growing interest in the development of person-fit statistics for tests with polytomous items. Some of the most popular person-fit statistics for such tests belong to the class of standardized person-fit statistics, <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mi>T</mi>\u0000 </mrow>\u0000 </semantics></math>, that is assumed to have a standard normal null distribution. However, this distribution only holds when (a) the true ability parameter is known and (b) an infinite number of items are available. In practice, both conditions are violated, and the quality of person-fit results is expected to deteriorate. In this paper, we propose three new corrections for <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mi>T</mi>\u0000 </mrow>\u0000 </semantics></math> that simultaneously account for the use of an estimated ability parameter and the use of a finite number of items. The three new corrections are direct extensions of those that were developed by Gorney et al. (<i>Psychometrika</i>, 2024, https://doi.org/10.1007/s11336-024-09960-x) for tests with only dichotomous items. Our simulation study reveals that the three new corrections tend to outperform not only the original statistic <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mi>T</mi>\u0000 </mrow>\u0000 </semantics></math> but also an existing correction for <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mi>T</mi>\u0000 </mrow>\u0000 </semantics></math> proposed by Sinharay (<i>Psychometrika</i>, 2016, 81, 992). Therefore, the new corrections appear to be promising tools for assessing person fit in tests with polytomous items.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"634-650"},"PeriodicalIF":1.5,"publicationDate":"2024-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12342","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140629021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Modelling motion energy in psychotherapy: A dynamical systems approach","authors":"Itai Dattner","doi":"10.1111/bmsp.12341","DOIUrl":"10.1111/bmsp.12341","url":null,"abstract":"<p>In this study we introduce an innovative mathematical and statistical framework for the analysis of motion energy dynamics in psychotherapy sessions. Our method combines motion energy dynamics with coupled linear ordinary differential equations and a measurement error model, contributing new clinical parameters to enhance psychotherapy research. Our approach transforms raw motion energy data into an interpretable account of therapist–patient interactions, providing novel insights into the dynamics of these interactions. A key aspect of our framework is the development of a new measure of synchrony between the motion energies of therapists and patients, which holds significant clinical and theoretical value in psychotherapy. The practical applicability and effectiveness of our modelling and estimation framework are demonstrated through the analysis of real session data. This work advances the quantitative analysis of motion dynamics in psychotherapy, offering important implications for future research and therapeutic practice.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"613-633"},"PeriodicalIF":1.5,"publicationDate":"2024-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12341","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140611390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Assessing quality of selection procedures: Lower bound of false positive rate as a function of inter-rater reliability","authors":"František Bartoš, Patrícia Martinková","doi":"10.1111/bmsp.12343","DOIUrl":"10.1111/bmsp.12343","url":null,"abstract":"<p>Inter-rater reliability (IRR) is one of the commonly used tools for assessing the quality of ratings from multiple raters. However, applicant selection procedures based on ratings from multiple raters usually result in a binary outcome; the applicant is either selected or not. This final outcome is not considered in IRR, which instead focuses on the ratings of the individual subjects or objects. We outline the connection between the ratings' measurement model (used for IRR) and a binary classification framework. We develop a simple way of approximating the probability of correctly selecting the best applicants which allows us to compute error probabilities of the selection procedure (i.e., false positive and false negative rate) or their lower bounds. We draw connections between the IRR and the binary classification metrics, showing that binary classification metrics depend solely on the IRR coefficient and proportion of selected applicants. We assess the performance of the approximation in a simulation study and apply it in an example comparing the reliability of multiple grant peer review selection procedures. We also discuss other possible uses of the explored connections in other contexts, such as educational testing, psychological assessment, and health-related measurement, and implement the computations in the R package <span>IRR2FPR</span>.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"651-671"},"PeriodicalIF":1.5,"publicationDate":"2024-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12343","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140592832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A comparison of different measures of the proportion of explained variance in multiply imputed data sets","authors":"Joost R. van Ginkel, Julian D. Karch","doi":"10.1111/bmsp.12344","DOIUrl":"10.1111/bmsp.12344","url":null,"abstract":"<p>The proportion of explained variance is an important statistic in multiple regression for determining how well the outcome variable is predicted by the predictors. Earlier research on 20 different estimators for the proportion of explained variance, including the exact Olkin–Pratt estimator and the Ezekiel estimator, showed that the exact Olkin–Pratt estimator produced unbiased estimates, and was recommended as a default estimator. In the current study, the same 20 estimators were studied in incomplete data, with missing data being treated using multiple imputation. In earlier research on the proportion of explained variance in multiply imputed data sets, an estimator called <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msubsup>\u0000 <mover>\u0000 <mi>R</mi>\u0000 <mo>̂</mo>\u0000 </mover>\u0000 <mi>SP</mi>\u0000 <mn>2</mn>\u0000 </msubsup>\u0000 </mrow>\u0000 </semantics></math> was shown to be the preferred pooled estimator for regular <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msup>\u0000 <mi>R</mi>\u0000 <mn>2</mn>\u0000 </msup>\u0000 </mrow>\u0000 </semantics></math>. For each of the 20 estimators in the current study, two pooled estimators were proposed: one where the estimator was the average across imputed data sets, and one where <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msubsup>\u0000 <mover>\u0000 <mi>R</mi>\u0000 <mo>̂</mo>\u0000 </mover>\u0000 <mi>SP</mi>\u0000 <mn>2</mn>\u0000 </msubsup>\u0000 </mrow>\u0000 </semantics></math> was used as input for the calculation of the specific estimator. Simulations showed that estimates based on <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msubsup>\u0000 <mover>\u0000 <mi>R</mi>\u0000 <mo>̂</mo>\u0000 </mover>\u0000 <mi>SP</mi>\u0000 <mn>2</mn>\u0000 </msubsup>\u0000 </mrow>\u0000 </semantics></math> performed best regarding bias and accuracy, and that the Ezekiel estimator was generally the least biased. However, none of the estimators were unbiased at all times, including the exact Olkin–Pratt estimator based on <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <msubsup>\u0000 <mover>\u0000 <mi>R</mi>\u0000 <mo>̂</mo>\u0000 </mover>\u0000 <mi>SP","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"672-693"},"PeriodicalIF":1.5,"publicationDate":"2024-04-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12344","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140592929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A two-step item bank calibration strategy based on 1-bit matrix completion for small-scale computerized adaptive testing","authors":"Yawei Shen, Shiyu Wang, Houping Xiao","doi":"10.1111/bmsp.12340","DOIUrl":"10.1111/bmsp.12340","url":null,"abstract":"<p>Computerized adaptive testing (CAT) is a widely embraced approach for delivering personalized educational assessments, tailoring each test to the real-time performance of individual examinees. Despite its potential advantages, CAT�s application in small-scale assessments has been limited due to the complexities associated with calibrating the item bank using sparse response data and small sample sizes. This study addresses these challenges by developing a two-step item bank calibration strategy that leverages the 1-bit matrix completion method in conjunction with two distinct incomplete pretesting designs. We introduce two novel 1-bit matrix completion-based imputation methods specifically designed to tackle the issues associated with item calibration in the presence of sparse response data and limited sample sizes. To demonstrate the effectiveness of these approaches, we conduct a comparative assessment against several established item parameter estimation methods capable of handling missing data. This evaluation is carried out through two sets of simulation studies, each featuring different pretesting designs, item bank structures, and sample sizes. Furthermore, we illustrate the practical application of the methods investigated, using empirical data collected from small-scale assessments.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"553-612"},"PeriodicalIF":1.5,"publicationDate":"2024-04-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12340","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140592809","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Shi-Fang Qiu, Jie Lei, Wai-Yin Poon, Man-Lai Tang, Ricky S. Wong, Ji-Ran Tao
{"title":"Sample size determination for interval estimation of the prevalence of a sensitive attribute under non-randomized response models","authors":"Shi-Fang Qiu, Jie Lei, Wai-Yin Poon, Man-Lai Tang, Ricky S. Wong, Ji-Ran Tao","doi":"10.1111/bmsp.12338","DOIUrl":"10.1111/bmsp.12338","url":null,"abstract":"<p>A sufficient number of participants should be included to adequately address the research interest in the surveys with sensitive questions. In this paper, sample size formulas/iterative algorithms are developed from the perspective of controlling the confidence interval width of the prevalence of a sensitive attribute under four non-randomized response models: the crosswise model, parallel model, Poisson item count technique model and negative binomial item count technique model. In contrast to the conventional approach for sample size determination, our sample size formulas/algorithms explicitly incorporate an assurance probability of controlling the width of a confidence interval within the pre-specified range. The performance of the proposed methods is evaluated with respect to the empirical coverage probability, empirical assurance probability and confidence width. Simulation results show that all formulas/algorithms are effective and hence are recommended for practical applications. A real example is used to illustrate the proposed methods.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"508-531"},"PeriodicalIF":1.5,"publicationDate":"2024-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139974734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sebastian Castro-Alvarez, Sandip Sinharay, Laura F. Bringmann, Rob R. Meijer, Jorge N. Tendeiro
{"title":"Assessment of fit of the time-varying dynamic partial credit model using the posterior predictive model checking method","authors":"Sebastian Castro-Alvarez, Sandip Sinharay, Laura F. Bringmann, Rob R. Meijer, Jorge N. Tendeiro","doi":"10.1111/bmsp.12339","DOIUrl":"10.1111/bmsp.12339","url":null,"abstract":"<p>Several new models based on item response theory have recently been suggested to analyse intensive longitudinal data. One of these new models is the time-varying dynamic partial credit model (TV-DPCM; Castro-Alvarez et al., <i>Multivariate Behavioral Research</i>, 2023, 1), which is a combination of the partial credit model and the time-varying autoregressive model. The model allows the study of the psychometric properties of the items and the modelling of nonlinear trends at the latent state level. However, there is a severe lack of tools to assess the fit of the TV-DPCM. In this paper, we propose and develop several test statistics and discrepancy measures based on the posterior predictive model checking (PPMC) method (PPMC; Rubin, <i>The Annals of Statistics</i>, 1984, 12, 1151) to assess the fit of the TV-DPCM. Simulated and empirical data are used to study the performance of and illustrate the effectiveness of the PPMC method.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"532-552"},"PeriodicalIF":1.5,"publicationDate":"2024-02-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139914100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"When and how to use set-exploratory structural equation modelling to test structural models: A tutorial using the R package lavaan","authors":"Herb Marsh, Abdullah Alamer","doi":"10.1111/bmsp.12336","DOIUrl":"10.1111/bmsp.12336","url":null,"abstract":"<p>Exploratory structural equation modelling (ESEM) is an alternative to the well-known method of confirmatory factor analysis (CFA). ESEM is mainly used to assess the quality of measurement models of common factors but can be efficiently extended to test structural models. However, ESEM may not be the best option in some model specifications, especially when structural models are involved, because the full flexibility of ESEM could result in technical difficulties in model estimation. Thus, set-ESEM was developed to accommodate the balance between full-ESEM and CFA. In the present paper, we show examples where set-ESEM should be used rather than full-ESEM. Rather than relying on a simulation study, we provide two applied examples using real data that are included in the OSF repository. Additionally, we provide the code needed to run set-ESEM in the free R package <i>lavaan</i> to make the paper practical. Set-ESEM structural models outperform their CFA-based counterparts in terms of goodness of fit and realistic factor correlation, and hence path coefficients in the two empirical examples. In several instances, effects that were non-significant (i.e., attenuated) in the CFA-based structural model become larger and significant in the set-ESEM structural model, suggesting that set-ESEM models may generate more accurate model parameters and, hence, lower Type II error rate.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"459-476"},"PeriodicalIF":1.5,"publicationDate":"2024-02-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12336","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139742778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Fast estimation of generalized linear latent variable models for performance and process data with ordinal, continuous, and count observed variables","authors":"Maoxin Zhang, Björn Andersson, Shaobo Jin","doi":"10.1111/bmsp.12337","DOIUrl":"10.1111/bmsp.12337","url":null,"abstract":"<p>Different data types often occur in psychological and educational measurement such as computer-based assessments that record performance and process data (e.g., response times and the number of actions). Modelling such data requires specific models for each data type and accommodating complex dependencies between multiple variables. Generalized linear latent variable models are suitable for modelling mixed data simultaneously, but estimation can be computationally demanding. A fast solution is to use Laplace approximations, but existing implementations of joint modelling of mixed data types are limited to ordinal and continuous data. To address this limitation, we derive an efficient estimation method that uses first- or second-order Laplace approximations to simultaneously model ordinal data, continuous data, and count data. We illustrate the approach with an example and conduct simulations to evaluate the performance of the method in terms of estimation efficiency, convergence, and parameter recovery. The results suggest that the second-order Laplace approximation achieves a higher convergence rate and produces accurate yet fast parameter estimates compared to the first-order Laplace approximation, while the time cost increases with higher model complexity. Additionally, models that consider the dependence of variables from the same stimulus fit the empirical data substantially better than models that disregarded the dependence.</p>","PeriodicalId":55322,"journal":{"name":"British Journal of Mathematical & Statistical Psychology","volume":"77 3","pages":"477-507"},"PeriodicalIF":1.5,"publicationDate":"2024-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/bmsp.12337","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139725087","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}