Parts2Whole:通过重构进行自监督对比学习

Ruibin Feng, Zongwei Zhou, Michael B Gotway, Jianming Liang
{"title":"Parts2Whole:通过重构进行自监督对比学习","authors":"Ruibin Feng, Zongwei Zhou, Michael B Gotway, Jianming Liang","doi":"10.1007/978-3-030-60548-3_9","DOIUrl":null,"url":null,"abstract":"<p><p>Contrastive representation learning is the state of the art in computer vision, but requires huge mini-batch sizes, special network design, or memory banks, making it unappealing for 3D medical imaging, while in 3D medical imaging, reconstruction-based self-supervised learning reaches a new height in performance, but lacks mechanisms to learn contrastive representation; therefore, this paper proposes a new framework for self-supervised contrastive learning via reconstruction, called Parts2Whole, because it exploits the <i>universal</i> and <i>intrinsic</i> part-whole relationship to learn contrastive representation without using contrastive loss: Reconstructing an image (whole) from its own parts compels the model to learn similar latent features for all its own parts, while reconstructing different images (wholes) from their respective parts forces the model to simultaneously push those parts belonging to different wholes farther apart from each other in the latent space; thereby the trained model is capable of distinguishing images. We have evaluated our Parts2Whole on five distinct imaging tasks covering both classification and segmentation, and compared it with four competing publicly available 3D pretrained models, showing that Parts2Whole significantly outperforms in two out of five tasks while achieves competitive performance on the rest three. This superior performance is attributable to the contrastive representations learned with Parts2Whole. Codes and pretrained models are available at github.com/JLiangLab/Parts2Whole.</p>","PeriodicalId":93545,"journal":{"name":"Domain adaptation and representation transfer, and distributed and collaborative learning : second MICCAI Workshop, DART 2020, and first MICCAI Workshop, DCL 2020, held in conjunction with MICCAI 2020, Lima, Peru, October 4-8, 2020, Pro...","volume":"73 1","pages":"85-95"},"PeriodicalIF":0.0000,"publicationDate":"2020-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9197952/pdf/","citationCount":"0","resultStr":"{\"title\":\"Parts2Whole: Self-supervised Contrastive Learning via Reconstruction.\",\"authors\":\"Ruibin Feng, Zongwei Zhou, Michael B Gotway, Jianming Liang\",\"doi\":\"10.1007/978-3-030-60548-3_9\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Contrastive representation learning is the state of the art in computer vision, but requires huge mini-batch sizes, special network design, or memory banks, making it unappealing for 3D medical imaging, while in 3D medical imaging, reconstruction-based self-supervised learning reaches a new height in performance, but lacks mechanisms to learn contrastive representation; therefore, this paper proposes a new framework for self-supervised contrastive learning via reconstruction, called Parts2Whole, because it exploits the <i>universal</i> and <i>intrinsic</i> part-whole relationship to learn contrastive representation without using contrastive loss: Reconstructing an image (whole) from its own parts compels the model to learn similar latent features for all its own parts, while reconstructing different images (wholes) from their respective parts forces the model to simultaneously push those parts belonging to different wholes farther apart from each other in the latent space; thereby the trained model is capable of distinguishing images. We have evaluated our Parts2Whole on five distinct imaging tasks covering both classification and segmentation, and compared it with four competing publicly available 3D pretrained models, showing that Parts2Whole significantly outperforms in two out of five tasks while achieves competitive performance on the rest three. This superior performance is attributable to the contrastive representations learned with Parts2Whole. Codes and pretrained models are available at github.com/JLiangLab/Parts2Whole.</p>\",\"PeriodicalId\":93545,\"journal\":{\"name\":\"Domain adaptation and representation transfer, and distributed and collaborative learning : second MICCAI Workshop, DART 2020, and first MICCAI Workshop, DCL 2020, held in conjunction with MICCAI 2020, Lima, Peru, October 4-8, 2020, Pro...\",\"volume\":\"73 1\",\"pages\":\"85-95\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9197952/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Domain adaptation and representation transfer, and distributed and collaborative learning : second MICCAI Workshop, DART 2020, and first MICCAI Workshop, DCL 2020, held in conjunction with MICCAI 2020, Lima, Peru, October 4-8, 2020, Pro...\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1007/978-3-030-60548-3_9\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2020/9/26 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Domain adaptation and representation transfer, and distributed and collaborative learning : second MICCAI Workshop, DART 2020, and first MICCAI Workshop, DCL 2020, held in conjunction with MICCAI 2020, Lima, Peru, October 4-8, 2020, Pro...","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/978-3-030-60548-3_9","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2020/9/26 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

对比度表征学习是计算机视觉领域的最新技术,但需要巨大的迷你批量、特殊的网络设计或内存库,因此对三维医学影像没有吸引力;而在三维医学影像中,基于重构的自我监督学习在性能上达到了一个新的高度,但缺乏学习对比度表征的机制;因此,本文提出了一种通过重构进行自我监督对比度学习的新框架,称为 "Parts2Whole",因为它利用了普遍的、内在的部分-整体关系,在不使用对比度损失的情况下学习对比度表征:从图像(整体)的各个部分重构图像(整体)会迫使模型为其自身的所有部分学习相似的潜在特征,而从不同的图像(整体)的各个部分重构图像(整体)则会迫使模型同时将属于不同整体的各个部分在潜在空间中推得更远;这样,训练有素的模型就能够区分图像。我们在五项不同的成像任务(包括分类和分割)中对 Parts2Whole 进行了评估,并将其与四种公开可用的三维预训练模型进行了比较,结果表明,Parts2Whole 在五项任务中的两项任务中的表现明显优于其他三项任务,并在其他三项任务中取得了具有竞争力的表现。这种优异的表现归功于 Parts2Whole 学习到的对比性表征。代码和预训练模型见 github.com/JLiangLab/Parts2Whole。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Parts2Whole: Self-supervised Contrastive Learning via Reconstruction.

Contrastive representation learning is the state of the art in computer vision, but requires huge mini-batch sizes, special network design, or memory banks, making it unappealing for 3D medical imaging, while in 3D medical imaging, reconstruction-based self-supervised learning reaches a new height in performance, but lacks mechanisms to learn contrastive representation; therefore, this paper proposes a new framework for self-supervised contrastive learning via reconstruction, called Parts2Whole, because it exploits the universal and intrinsic part-whole relationship to learn contrastive representation without using contrastive loss: Reconstructing an image (whole) from its own parts compels the model to learn similar latent features for all its own parts, while reconstructing different images (wholes) from their respective parts forces the model to simultaneously push those parts belonging to different wholes farther apart from each other in the latent space; thereby the trained model is capable of distinguishing images. We have evaluated our Parts2Whole on five distinct imaging tasks covering both classification and segmentation, and compared it with four competing publicly available 3D pretrained models, showing that Parts2Whole significantly outperforms in two out of five tasks while achieves competitive performance on the rest three. This superior performance is attributable to the contrastive representations learned with Parts2Whole. Codes and pretrained models are available at github.com/JLiangLab/Parts2Whole.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信