{"title":"不适合所有人的量身定制的拟合:诊断研究中的阈值过拟合问题。","authors":"Javier Arredondo Montero","doi":"10.1515/dx-2025-0096","DOIUrl":null,"url":null,"abstract":"<p><strong>Objectives: </strong>To critically examine the phenomenon of threshold overfitting in diagnostic accuracy research and evaluate its methodological implications through a structured review of relevant literature.</p><p><strong>Methods: </strong>This article presents a narrative and critical review of methodological studies and reporting guidelines related to threshold selection in diagnostic test accuracy. It focuses on the misuse of <i>post hoc</i> thresholds, the misapplication of bias assessment tools such as QUADAS-2, and the frequent absence of independent validation. In addition to identifying these structural flaws, the article proposes a set of five concrete safeguards - ranging from transparent reporting to rigorous risk of bias classification - designed to mitigate threshold-related bias in future diagnostic studies.</p><p><strong>Results: </strong>Thresholds are frequently derived and evaluated within the same dataset, inflating sensitivity and specificity estimates. This overfitting is seldom acknowledged and is often misclassified as low risk of bias. QUADAS-2 is frequently misapplied, with reviewers mistaking the mere presence of a threshold for proper pre-specification. The article identifies five key safeguards to mitigate this bias: (1) clear declaration of pre-specification, (2) justification of threshold choice, (3) independent validation, (4) full performance reporting across thresholds, and (5) rigorous application of bias assessment tools.</p><p><strong>Conclusions: </strong>Threshold overfitting remains an underrecognized but methodologically critical source of bias in diagnostic accuracy studies. Addressing it requires more than awareness - it demands transparent reporting, proper validation, and stricter adherence to methodological standards.</p>","PeriodicalId":11273,"journal":{"name":"Diagnosis","volume":" ","pages":""},"PeriodicalIF":2.0000,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A tailored fit that doesn't fit all: the problem of threshold overfitting in diagnostic studies.\",\"authors\":\"Javier Arredondo Montero\",\"doi\":\"10.1515/dx-2025-0096\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Objectives: </strong>To critically examine the phenomenon of threshold overfitting in diagnostic accuracy research and evaluate its methodological implications through a structured review of relevant literature.</p><p><strong>Methods: </strong>This article presents a narrative and critical review of methodological studies and reporting guidelines related to threshold selection in diagnostic test accuracy. It focuses on the misuse of <i>post hoc</i> thresholds, the misapplication of bias assessment tools such as QUADAS-2, and the frequent absence of independent validation. In addition to identifying these structural flaws, the article proposes a set of five concrete safeguards - ranging from transparent reporting to rigorous risk of bias classification - designed to mitigate threshold-related bias in future diagnostic studies.</p><p><strong>Results: </strong>Thresholds are frequently derived and evaluated within the same dataset, inflating sensitivity and specificity estimates. This overfitting is seldom acknowledged and is often misclassified as low risk of bias. QUADAS-2 is frequently misapplied, with reviewers mistaking the mere presence of a threshold for proper pre-specification. The article identifies five key safeguards to mitigate this bias: (1) clear declaration of pre-specification, (2) justification of threshold choice, (3) independent validation, (4) full performance reporting across thresholds, and (5) rigorous application of bias assessment tools.</p><p><strong>Conclusions: </strong>Threshold overfitting remains an underrecognized but methodologically critical source of bias in diagnostic accuracy studies. Addressing it requires more than awareness - it demands transparent reporting, proper validation, and stricter adherence to methodological standards.</p>\",\"PeriodicalId\":11273,\"journal\":{\"name\":\"Diagnosis\",\"volume\":\" \",\"pages\":\"\"},\"PeriodicalIF\":2.0000,\"publicationDate\":\"2025-09-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Diagnosis\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1515/dx-2025-0096\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"MEDICINE, GENERAL & INTERNAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Diagnosis","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1515/dx-2025-0096","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MEDICINE, GENERAL & INTERNAL","Score":null,"Total":0}
A tailored fit that doesn't fit all: the problem of threshold overfitting in diagnostic studies.
Objectives: To critically examine the phenomenon of threshold overfitting in diagnostic accuracy research and evaluate its methodological implications through a structured review of relevant literature.
Methods: This article presents a narrative and critical review of methodological studies and reporting guidelines related to threshold selection in diagnostic test accuracy. It focuses on the misuse of post hoc thresholds, the misapplication of bias assessment tools such as QUADAS-2, and the frequent absence of independent validation. In addition to identifying these structural flaws, the article proposes a set of five concrete safeguards - ranging from transparent reporting to rigorous risk of bias classification - designed to mitigate threshold-related bias in future diagnostic studies.
Results: Thresholds are frequently derived and evaluated within the same dataset, inflating sensitivity and specificity estimates. This overfitting is seldom acknowledged and is often misclassified as low risk of bias. QUADAS-2 is frequently misapplied, with reviewers mistaking the mere presence of a threshold for proper pre-specification. The article identifies five key safeguards to mitigate this bias: (1) clear declaration of pre-specification, (2) justification of threshold choice, (3) independent validation, (4) full performance reporting across thresholds, and (5) rigorous application of bias assessment tools.
Conclusions: Threshold overfitting remains an underrecognized but methodologically critical source of bias in diagnostic accuracy studies. Addressing it requires more than awareness - it demands transparent reporting, proper validation, and stricter adherence to methodological standards.
期刊介绍:
Diagnosis focuses on how diagnosis can be advanced, how it is taught, and how and why it can fail, leading to diagnostic errors. The journal welcomes both fundamental and applied works, improvement initiatives, opinions, and debates to encourage new thinking on improving this critical aspect of healthcare quality. Topics: -Factors that promote diagnostic quality and safety -Clinical reasoning -Diagnostic errors in medicine -The factors that contribute to diagnostic error: human factors, cognitive issues, and system-related breakdowns -Improving the value of diagnosis – eliminating waste and unnecessary testing -How culture and removing blame promote awareness of diagnostic errors -Training and education related to clinical reasoning and diagnostic skills -Advances in laboratory testing and imaging that improve diagnostic capability -Local, national and international initiatives to reduce diagnostic error