离散-连续混合反应的项目反应理论模型比较研究。

IF 2.8 3区 心理学 Q1 PSYCHOLOGY, MULTIDISCIPLINARY
Cengiz Zopluoglu, J R Lockwood
{"title":"离散-连续混合反应的项目反应理论模型比较研究。","authors":"Cengiz Zopluoglu, J R Lockwood","doi":"10.3390/jintelligence12030026","DOIUrl":null,"url":null,"abstract":"<p><p>Language proficiency assessments are pivotal in educational and professional decision-making. With the integration of AI-driven technologies, these assessments can more frequently use item types, such as dictation tasks, producing response features with a mixture of discrete and continuous distributions. This study evaluates novel measurement models tailored to these unique response features. Specifically, we evaluated the performance of the zero-and-one-inflated extensions of the Beta, Simplex, and Samejima's Continuous item response models and incorporated collateral information into the estimation using latent regression. Our findings highlight that while all models provided highly correlated results regarding item and person parameters, the Beta item response model showcased superior out-of-sample predictive accuracy. However, a significant challenge was the absence of established benchmarks for evaluating model and item fit for these novel item response models. There is a need for further research to establish benchmarks for evaluating the fit of these innovative models to ensure their reliability and validity in real-world applications.</p>","PeriodicalId":52279,"journal":{"name":"Journal of Intelligence","volume":"12 3","pages":""},"PeriodicalIF":2.8000,"publicationDate":"2024-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10970766/pdf/","citationCount":"0","resultStr":"{\"title\":\"A Comparative Study of Item Response Theory Models for Mixed Discrete-Continuous Responses.\",\"authors\":\"Cengiz Zopluoglu, J R Lockwood\",\"doi\":\"10.3390/jintelligence12030026\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Language proficiency assessments are pivotal in educational and professional decision-making. With the integration of AI-driven technologies, these assessments can more frequently use item types, such as dictation tasks, producing response features with a mixture of discrete and continuous distributions. This study evaluates novel measurement models tailored to these unique response features. Specifically, we evaluated the performance of the zero-and-one-inflated extensions of the Beta, Simplex, and Samejima's Continuous item response models and incorporated collateral information into the estimation using latent regression. Our findings highlight that while all models provided highly correlated results regarding item and person parameters, the Beta item response model showcased superior out-of-sample predictive accuracy. However, a significant challenge was the absence of established benchmarks for evaluating model and item fit for these novel item response models. There is a need for further research to establish benchmarks for evaluating the fit of these innovative models to ensure their reliability and validity in real-world applications.</p>\",\"PeriodicalId\":52279,\"journal\":{\"name\":\"Journal of Intelligence\",\"volume\":\"12 3\",\"pages\":\"\"},\"PeriodicalIF\":2.8000,\"publicationDate\":\"2024-02-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10970766/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Intelligence\",\"FirstCategoryId\":\"102\",\"ListUrlMain\":\"https://doi.org/10.3390/jintelligence12030026\",\"RegionNum\":3,\"RegionCategory\":\"心理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"PSYCHOLOGY, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Intelligence","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.3390/jintelligence12030026","RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHOLOGY, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

摘要

语言能力评估在教育和专业决策中举足轻重。随着人工智能技术的融入,这些评估可以更频繁地使用听写任务等项目类型,从而产生离散和连续混合分布的反应特征。本研究评估了针对这些独特应答特征而定制的新型测量模型。具体来说,我们评估了 Beta、Simplex 和 Samejima 连续项目反应模型的零膨胀和一膨胀扩展模型的性能,并使用潜在回归将附带信息纳入估计。我们的研究结果表明,虽然所有模型都提供了与项目和个人参数高度相关的结果,但 Beta 项目反应模型显示出更高的样本外预测准确性。然而,一个重大的挑战是缺乏既定的基准来评估这些新型项目反应模型的模型和项目拟合度。有必要开展进一步的研究,建立评估这些创新模型拟合度的基准,以确保它们在实际应用中的可靠性和有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A Comparative Study of Item Response Theory Models for Mixed Discrete-Continuous Responses.

Language proficiency assessments are pivotal in educational and professional decision-making. With the integration of AI-driven technologies, these assessments can more frequently use item types, such as dictation tasks, producing response features with a mixture of discrete and continuous distributions. This study evaluates novel measurement models tailored to these unique response features. Specifically, we evaluated the performance of the zero-and-one-inflated extensions of the Beta, Simplex, and Samejima's Continuous item response models and incorporated collateral information into the estimation using latent regression. Our findings highlight that while all models provided highly correlated results regarding item and person parameters, the Beta item response model showcased superior out-of-sample predictive accuracy. However, a significant challenge was the absence of established benchmarks for evaluating model and item fit for these novel item response models. There is a need for further research to establish benchmarks for evaluating the fit of these innovative models to ensure their reliability and validity in real-world applications.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Journal of Intelligence
Journal of Intelligence Social Sciences-Education
CiteScore
2.80
自引率
17.10%
发文量
0
审稿时长
11 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信