{"title":"大ACR MRI幻像低对比度目标可检测性测试自动分析的统计方法","authors":"Ali M. Golestani, Julia M. Gee","doi":"10.1002/acm2.70173","DOIUrl":null,"url":null,"abstract":"<div>\n \n \n <section>\n \n <h3> Background</h3>\n \n <p>Regular quality control checks are essential to ensure the quality of MRI systems. The American College of Radiology (ACR) has developed a standardized large phantom test protocol for this purpose. However, the ACR protocol recommends manual measurements, which are time-consuming, labor-intensive, and prone to variability, impacting accuracy and reproducibility. Although some aspects of the ACR evaluation have been automated or semi-automated, tests like low-contrast object detectability (LCOD), remain challenging to automate. LCOD involves assessing the visibility of objects at various contrast levels.</p>\n </section>\n \n <section>\n \n <h3> Purpose</h3>\n \n <p>The purpose of this research is to propose and evaluate an automated approach for LCOD testing in MRI.</p>\n </section>\n \n <section>\n \n <h3> Methods</h3>\n \n <p>The automated Python code generates a one-dimensional profile of image intensities along radial paths from the center of the contrast disk. These profiles are compared to templates created from the disc's geometric information using general linear model statistical tests. A total of 80 image volumes (40 T1- and 40 T2-weighted) were assessed twice by two human evaluators and the proposed Python code.</p>\n </section>\n \n <section>\n \n <h3> Results</h3>\n \n <p>Human raters showed intra-rater variability (Cohen's Kappa 0.941, 0.962), while the Python code exhibited perfect intra-rater agreement. Inter-rater agreement between the code and humans was comparable to human-to-human agreement (Cohen's Kappa 0.878 between the two human raters vs. 0.945, and 0.783 between the code and human raters). A stress test revealed both human raters and the code assigned higher scores to lower bandwidth images and lower scores to higher bandwidth images.</p>\n </section>\n \n <section>\n \n <h3> Conclusion</h3>\n \n <p>The proposed automated method eliminates intra-rater variability and achieves strong inter-rater agreement with human raters. These findings suggest the method is reliable and suitable for clinical settings, showing high concordance with human assessments.</p>\n </section>\n </div>","PeriodicalId":14989,"journal":{"name":"Journal of Applied Clinical Medical Physics","volume":"26 7","pages":""},"PeriodicalIF":2.2000,"publicationDate":"2025-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/acm2.70173","citationCount":"0","resultStr":"{\"title\":\"A statistical approach to automated analysis of the low-contrast object detectability test for the large ACR MRI phantom\",\"authors\":\"Ali M. Golestani, Julia M. Gee\",\"doi\":\"10.1002/acm2.70173\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n \\n <section>\\n \\n <h3> Background</h3>\\n \\n <p>Regular quality control checks are essential to ensure the quality of MRI systems. The American College of Radiology (ACR) has developed a standardized large phantom test protocol for this purpose. However, the ACR protocol recommends manual measurements, which are time-consuming, labor-intensive, and prone to variability, impacting accuracy and reproducibility. Although some aspects of the ACR evaluation have been automated or semi-automated, tests like low-contrast object detectability (LCOD), remain challenging to automate. LCOD involves assessing the visibility of objects at various contrast levels.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Purpose</h3>\\n \\n <p>The purpose of this research is to propose and evaluate an automated approach for LCOD testing in MRI.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Methods</h3>\\n \\n <p>The automated Python code generates a one-dimensional profile of image intensities along radial paths from the center of the contrast disk. These profiles are compared to templates created from the disc's geometric information using general linear model statistical tests. A total of 80 image volumes (40 T1- and 40 T2-weighted) were assessed twice by two human evaluators and the proposed Python code.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Results</h3>\\n \\n <p>Human raters showed intra-rater variability (Cohen's Kappa 0.941, 0.962), while the Python code exhibited perfect intra-rater agreement. Inter-rater agreement between the code and humans was comparable to human-to-human agreement (Cohen's Kappa 0.878 between the two human raters vs. 0.945, and 0.783 between the code and human raters). A stress test revealed both human raters and the code assigned higher scores to lower bandwidth images and lower scores to higher bandwidth images.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Conclusion</h3>\\n \\n <p>The proposed automated method eliminates intra-rater variability and achieves strong inter-rater agreement with human raters. These findings suggest the method is reliable and suitable for clinical settings, showing high concordance with human assessments.</p>\\n </section>\\n </div>\",\"PeriodicalId\":14989,\"journal\":{\"name\":\"Journal of Applied Clinical Medical Physics\",\"volume\":\"26 7\",\"pages\":\"\"},\"PeriodicalIF\":2.2000,\"publicationDate\":\"2025-07-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1002/acm2.70173\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Applied Clinical Medical Physics\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/acm2.70173\",\"RegionNum\":4,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Applied Clinical Medical Physics","FirstCategoryId":"3","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/acm2.70173","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
A statistical approach to automated analysis of the low-contrast object detectability test for the large ACR MRI phantom
Background
Regular quality control checks are essential to ensure the quality of MRI systems. The American College of Radiology (ACR) has developed a standardized large phantom test protocol for this purpose. However, the ACR protocol recommends manual measurements, which are time-consuming, labor-intensive, and prone to variability, impacting accuracy and reproducibility. Although some aspects of the ACR evaluation have been automated or semi-automated, tests like low-contrast object detectability (LCOD), remain challenging to automate. LCOD involves assessing the visibility of objects at various contrast levels.
Purpose
The purpose of this research is to propose and evaluate an automated approach for LCOD testing in MRI.
Methods
The automated Python code generates a one-dimensional profile of image intensities along radial paths from the center of the contrast disk. These profiles are compared to templates created from the disc's geometric information using general linear model statistical tests. A total of 80 image volumes (40 T1- and 40 T2-weighted) were assessed twice by two human evaluators and the proposed Python code.
Results
Human raters showed intra-rater variability (Cohen's Kappa 0.941, 0.962), while the Python code exhibited perfect intra-rater agreement. Inter-rater agreement between the code and humans was comparable to human-to-human agreement (Cohen's Kappa 0.878 between the two human raters vs. 0.945, and 0.783 between the code and human raters). A stress test revealed both human raters and the code assigned higher scores to lower bandwidth images and lower scores to higher bandwidth images.
Conclusion
The proposed automated method eliminates intra-rater variability and achieves strong inter-rater agreement with human raters. These findings suggest the method is reliable and suitable for clinical settings, showing high concordance with human assessments.
期刊介绍:
Journal of Applied Clinical Medical Physics is an international Open Access publication dedicated to clinical medical physics. JACMP welcomes original contributions dealing with all aspects of medical physics from scientists working in the clinical medical physics around the world. JACMP accepts only online submission.
JACMP will publish:
-Original Contributions: Peer-reviewed, investigations that represent new and significant contributions to the field. Recommended word count: up to 7500.
-Review Articles: Reviews of major areas or sub-areas in the field of clinical medical physics. These articles may be of any length and are peer reviewed.
-Technical Notes: These should be no longer than 3000 words, including key references.
-Letters to the Editor: Comments on papers published in JACMP or on any other matters of interest to clinical medical physics. These should not be more than 1250 (including the literature) and their publication is only based on the decision of the editor, who occasionally asks experts on the merit of the contents.
-Book Reviews: The editorial office solicits Book Reviews.
-Announcements of Forthcoming Meetings: The Editor may provide notice of forthcoming meetings, course offerings, and other events relevant to clinical medical physics.
-Parallel Opposed Editorial: We welcome topics relevant to clinical practice and medical physics profession. The contents can be controversial debate or opposed aspects of an issue. One author argues for the position and the other against. Each side of the debate contains an opening statement up to 800 words, followed by a rebuttal up to 500 words. Readers interested in participating in this series should contact the moderator with a proposed title and a short description of the topic