乳房x光检查中的PGMI评估:人工智能软件与人类读者

IF 2.8 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING
T. Santner , C. Ruppert , S. Gianolini , J.-G. Stalheim , S. Frei , M. Hondl , V. Fröhlich , S. Hofvind , G. Widmann
{"title":"乳房x光检查中的PGMI评估:人工智能软件与人类读者","authors":"T. Santner ,&nbsp;C. Ruppert ,&nbsp;S. Gianolini ,&nbsp;J.-G. Stalheim ,&nbsp;S. Frei ,&nbsp;M. Hondl ,&nbsp;V. Fröhlich ,&nbsp;S. Hofvind ,&nbsp;G. Widmann","doi":"10.1016/j.radi.2025.103017","DOIUrl":null,"url":null,"abstract":"<div><h3>Introduction</h3><div>The aim of this study was to evaluate human inter-reader agreement of parameters included in PGMI (perfect-good-moderate-inadequate) classification of screening mammograms and explore the role of artificial intelligence (AI) as an alternative reader.</div></div><div><h3>Methods</h3><div>Five radiographers from three European countries independently performed a PGMI assessment of 520 anonymized mammography screening examinations randomly selected from representative subsets from 13 imaging centres within two European countries. As a sixth reader, a dedicated AI software was used. Accuracy, Cohen's Kappa, and confusion matrices were calculated to compare the predictions of the software against the individual assessment of the readers, as well as potential discrepancies between them. A questionnaire and a personality test were used to better understand the decision-making processes of the human readers.</div></div><div><h3>Results</h3><div>Significant inter-reader variability among human readers with poor to moderate agreement (κ = −0.018 to κ = 0.41) was observed, with some showing more homogenous interpretations of single features and overall quality than others. In comparison, the software surpassed human inter-reader agreement in detecting glandular tissue cuts, mammilla deviation, pectoral muscle detection, and pectoral angle measurement, while remaining features and overall image quality exhibited comparable performance to human assessment.</div></div><div><h3>Conclusion</h3><div>Notably, human inter-reader disagreement of PGMI assessment in mammography is considerably high. AI software may already reliably categorize quality. Its potential for standardization and immediate feedback to achieve and monitor high levels of quality in screening programs needs further attention and should be included in future approaches.</div></div><div><h3>Implications for practice</h3><div>AI has promising potential for automated assessment of diagnostic image quality. Faster, more representative and more objective feedback may support radiographers in their quality management processes. Direct transformation of common PGMI workflows into an AI algorithm could be challenging.</div></div>","PeriodicalId":47416,"journal":{"name":"Radiography","volume":"31 5","pages":"Article 103017"},"PeriodicalIF":2.8000,"publicationDate":"2025-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"PGMI assessment in mammography: AI software versus human readers\",\"authors\":\"T. Santner ,&nbsp;C. Ruppert ,&nbsp;S. Gianolini ,&nbsp;J.-G. Stalheim ,&nbsp;S. Frei ,&nbsp;M. Hondl ,&nbsp;V. Fröhlich ,&nbsp;S. Hofvind ,&nbsp;G. Widmann\",\"doi\":\"10.1016/j.radi.2025.103017\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><h3>Introduction</h3><div>The aim of this study was to evaluate human inter-reader agreement of parameters included in PGMI (perfect-good-moderate-inadequate) classification of screening mammograms and explore the role of artificial intelligence (AI) as an alternative reader.</div></div><div><h3>Methods</h3><div>Five radiographers from three European countries independently performed a PGMI assessment of 520 anonymized mammography screening examinations randomly selected from representative subsets from 13 imaging centres within two European countries. As a sixth reader, a dedicated AI software was used. Accuracy, Cohen's Kappa, and confusion matrices were calculated to compare the predictions of the software against the individual assessment of the readers, as well as potential discrepancies between them. A questionnaire and a personality test were used to better understand the decision-making processes of the human readers.</div></div><div><h3>Results</h3><div>Significant inter-reader variability among human readers with poor to moderate agreement (κ = −0.018 to κ = 0.41) was observed, with some showing more homogenous interpretations of single features and overall quality than others. In comparison, the software surpassed human inter-reader agreement in detecting glandular tissue cuts, mammilla deviation, pectoral muscle detection, and pectoral angle measurement, while remaining features and overall image quality exhibited comparable performance to human assessment.</div></div><div><h3>Conclusion</h3><div>Notably, human inter-reader disagreement of PGMI assessment in mammography is considerably high. AI software may already reliably categorize quality. Its potential for standardization and immediate feedback to achieve and monitor high levels of quality in screening programs needs further attention and should be included in future approaches.</div></div><div><h3>Implications for practice</h3><div>AI has promising potential for automated assessment of diagnostic image quality. Faster, more representative and more objective feedback may support radiographers in their quality management processes. Direct transformation of common PGMI workflows into an AI algorithm could be challenging.</div></div>\",\"PeriodicalId\":47416,\"journal\":{\"name\":\"Radiography\",\"volume\":\"31 5\",\"pages\":\"Article 103017\"},\"PeriodicalIF\":2.8000,\"publicationDate\":\"2025-07-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Radiography\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1078817425001610\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Radiography","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1078817425001610","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0

摘要

本研究的目的是评估筛查乳房x线照片PGMI(完美-良好-中等-不充分)分类中包含的参数的人类解读者间一致性,并探讨人工智能(AI)作为替代解读者的作用。方法来自3个欧洲国家的5名放射技师对从2个欧洲国家13个影像中心的代表性亚群中随机选择的520例匿名乳房x线摄影筛查检查进行了PGMI评估。作为第六个阅读器,使用了专用的人工智能软件。计算准确性、Cohen’s Kappa和混淆矩阵,以比较软件的预测与读者的个人评估,以及它们之间的潜在差异。为了更好地理解人类读者的决策过程,研究人员使用了问卷调查和性格测试。结果在一致性较差至中等的人类读者(κ = - 0.018至κ = 0.41)中观察到显着的读者间差异,其中一些读者对单个特征和整体质量的解释比其他读者更均匀。相比之下,该软件在检测腺组织切割、乳房偏离、胸肌检测和胸角测量方面超过了人类读者间协议,而其余特征和整体图像质量表现出与人类评估相当的性能。结论乳房x线摄影中PGMI评估的人类读者间歧异率相当高。人工智能软件可能已经对质量进行了可靠的分类。它在标准化和即时反馈方面的潜力,以实现和监测筛查项目的高质量,需要进一步关注,并应纳入未来的方法。人工智能在诊断图像质量的自动评估方面具有很大的潜力。更快、更有代表性和更客观的反馈可以帮助放射技师进行质量管理。将常见的PGMI工作流直接转换为人工智能算法可能具有挑战性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
PGMI assessment in mammography: AI software versus human readers

Introduction

The aim of this study was to evaluate human inter-reader agreement of parameters included in PGMI (perfect-good-moderate-inadequate) classification of screening mammograms and explore the role of artificial intelligence (AI) as an alternative reader.

Methods

Five radiographers from three European countries independently performed a PGMI assessment of 520 anonymized mammography screening examinations randomly selected from representative subsets from 13 imaging centres within two European countries. As a sixth reader, a dedicated AI software was used. Accuracy, Cohen's Kappa, and confusion matrices were calculated to compare the predictions of the software against the individual assessment of the readers, as well as potential discrepancies between them. A questionnaire and a personality test were used to better understand the decision-making processes of the human readers.

Results

Significant inter-reader variability among human readers with poor to moderate agreement (κ = −0.018 to κ = 0.41) was observed, with some showing more homogenous interpretations of single features and overall quality than others. In comparison, the software surpassed human inter-reader agreement in detecting glandular tissue cuts, mammilla deviation, pectoral muscle detection, and pectoral angle measurement, while remaining features and overall image quality exhibited comparable performance to human assessment.

Conclusion

Notably, human inter-reader disagreement of PGMI assessment in mammography is considerably high. AI software may already reliably categorize quality. Its potential for standardization and immediate feedback to achieve and monitor high levels of quality in screening programs needs further attention and should be included in future approaches.

Implications for practice

AI has promising potential for automated assessment of diagnostic image quality. Faster, more representative and more objective feedback may support radiographers in their quality management processes. Direct transformation of common PGMI workflows into an AI algorithm could be challenging.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Radiography
Radiography RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING-
CiteScore
4.70
自引率
34.60%
发文量
169
审稿时长
63 days
期刊介绍: Radiography is an International, English language, peer-reviewed journal of diagnostic imaging and radiation therapy. Radiography is the official professional journal of the College of Radiographers and is published quarterly. Radiography aims to publish the highest quality material, both clinical and scientific, on all aspects of diagnostic imaging and radiation therapy and oncology.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信