Self-Supervised Learning for Accurate Liver View Classification in Ultrasound Images with Minimal Labeled Data

Dr. Abder-Rahman Ali, A. Samir, Peng Guo
{"title":"Self-Supervised Learning for Accurate Liver View Classification in Ultrasound Images with Minimal Labeled Data","authors":"Dr. Abder-Rahman Ali, A. Samir, Peng Guo","doi":"10.1109/CVPRW59228.2023.00310","DOIUrl":null,"url":null,"abstract":"Conventional B-mode \"grey scale\" medical ultrasound and shear wave elastography (SWE) are widely used for chronic liver disease diagnosis and risk stratification. Liver disease is very common and is clinically and socially important. As a result, multiple medical device manufacturers have proposed or developed AI systems for ultrasound image analysis. However, many abdominal ultrasound images do not include views of the liver, necessitating manual data curation for model development. To optimize the efficiency of real-time processing, a pre-processing liver view detection step is necessary before feeding the image to the AI system. Deep learning techniques have shown great promise for image classification, yet labeling large datasets for training classification models is timeconsuming and expensive. In this paper, we present a selfsupervised learning method for image classification that utilizes a large set of unlabeled abdominal ultrasound images to learn image representations. These representations are then applied on the downstream task of liver view classification, resulting in efficient classification and alleviation of the labeling burden. In comparison to two state-of-the-art (SOTA) models, ResNet-18 and MLP-Mixer, when trained for 100 epochs the proposed SimCLR+LR approach demonstrated outstanding performance when only labeling \"one\" image per class, achieving an accuracy similar to MLP-Mixer (86%) and outperforming the performance of ResNet-18 (70.2%), when trained on 854 (with liver: 495, without liver: 359) B-mode images. When trained on the whole dataset for 1000 epochs, SimCLR+LR and ResNet-18 achieved an accuracy of 98.7% and 79.3%, respectively. These findings highlight the potential of the SimCLR+LR approach as a superior alternative to traditional supervised learning methods for liver view classification. Our proposed method has the ability to reduce both the time and cost associated with data labeling, as it eliminates the need for human labor (i.e., SOTA performance achieved with only a small amount of labeled data). The approach could also be advantageous in scenarios where a subset of images with a particular organ needs to be extracted from a large dataset that includes images of various organs.","PeriodicalId":355438,"journal":{"name":"2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CVPRW59228.2023.00310","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Conventional B-mode "grey scale" medical ultrasound and shear wave elastography (SWE) are widely used for chronic liver disease diagnosis and risk stratification. Liver disease is very common and is clinically and socially important. As a result, multiple medical device manufacturers have proposed or developed AI systems for ultrasound image analysis. However, many abdominal ultrasound images do not include views of the liver, necessitating manual data curation for model development. To optimize the efficiency of real-time processing, a pre-processing liver view detection step is necessary before feeding the image to the AI system. Deep learning techniques have shown great promise for image classification, yet labeling large datasets for training classification models is timeconsuming and expensive. In this paper, we present a selfsupervised learning method for image classification that utilizes a large set of unlabeled abdominal ultrasound images to learn image representations. These representations are then applied on the downstream task of liver view classification, resulting in efficient classification and alleviation of the labeling burden. In comparison to two state-of-the-art (SOTA) models, ResNet-18 and MLP-Mixer, when trained for 100 epochs the proposed SimCLR+LR approach demonstrated outstanding performance when only labeling "one" image per class, achieving an accuracy similar to MLP-Mixer (86%) and outperforming the performance of ResNet-18 (70.2%), when trained on 854 (with liver: 495, without liver: 359) B-mode images. When trained on the whole dataset for 1000 epochs, SimCLR+LR and ResNet-18 achieved an accuracy of 98.7% and 79.3%, respectively. These findings highlight the potential of the SimCLR+LR approach as a superior alternative to traditional supervised learning methods for liver view classification. Our proposed method has the ability to reduce both the time and cost associated with data labeling, as it eliminates the need for human labor (i.e., SOTA performance achieved with only a small amount of labeled data). The approach could also be advantageous in scenarios where a subset of images with a particular organ needs to be extracted from a large dataset that includes images of various organs.
基于最小标记数据的超声图像肝脏视图精确分类的自监督学习
常规b超“灰阶”医学超声和横波弹性成像(SWE)被广泛用于慢性肝病的诊断和危险分层。肝病是一种非常常见的疾病,具有重要的临床和社会意义。因此,多家医疗设备制造商提出或开发了用于超声图像分析的人工智能系统。然而,许多腹部超声图像不包括肝脏的视图,需要手工数据管理模型开发。为了优化实时处理的效率,在将图像输入人工智能系统之前,需要预处理肝脏视图检测步骤。深度学习技术在图像分类方面显示出巨大的前景,然而为训练分类模型标记大型数据集是耗时且昂贵的。在本文中,我们提出了一种用于图像分类的自监督学习方法,该方法利用大量未标记的腹部超声图像来学习图像表示。然后将这些表示应用于肝视图分类的下游任务,从而实现有效的分类并减轻标记负担。与两种最先进的(SOTA)模型ResNet-18和MLP-Mixer相比,当训练100次时,所提出的SimCLR+LR方法在每个类别只标记“一个”图像时表现出出色的性能,达到与MLP-Mixer相似的精度(86%),并且在854(含肝脏:495,无肝脏:359)b模式图像上训练时优于ResNet-18(70.2%)。在整个数据集上训练1000次时,SimCLR+LR和ResNet-18的准确率分别达到98.7%和79.3%。这些发现突出了SimCLR+LR方法作为传统监督学习方法的更好替代方案的潜力。我们提出的方法能够减少与数据标记相关的时间和成本,因为它消除了对人工劳动的需求(即,仅使用少量标记数据即可实现SOTA性能)。在需要从包含各种器官图像的大型数据集中提取具有特定器官的图像子集的情况下,该方法也可能是有利的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信