Lucas Farndale, Chris Walsh, Robert Insall, Ke Yuan
{"title":"合成特权信息加强医学图像表征学习","authors":"Lucas Farndale, Chris Walsh, Robert Insall, Ke Yuan","doi":"arxiv-2403.05220","DOIUrl":null,"url":null,"abstract":"Multimodal self-supervised representation learning has consistently proven to\nbe a highly effective method in medical image analysis, offering strong task\nperformance and producing biologically informed insights. However, these\nmethods heavily rely on large, paired datasets, which is prohibitive for their\nuse in scenarios where paired data does not exist, or there is only a small\namount available. In contrast, image generation methods can work well on very\nsmall datasets, and can find mappings between unpaired datasets, meaning an\neffectively unlimited amount of paired synthetic data can be generated. In this\nwork, we demonstrate that representation learning can be significantly improved\nby synthetically generating paired information, both compared to training on\neither single-modality (up to 4.4x error reduction) or authentic multi-modal\npaired datasets (up to 5.6x error reduction).","PeriodicalId":501572,"journal":{"name":"arXiv - QuanBio - Tissues and Organs","volume":"91 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Synthetic Privileged Information Enhances Medical Image Representation Learning\",\"authors\":\"Lucas Farndale, Chris Walsh, Robert Insall, Ke Yuan\",\"doi\":\"arxiv-2403.05220\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multimodal self-supervised representation learning has consistently proven to\\nbe a highly effective method in medical image analysis, offering strong task\\nperformance and producing biologically informed insights. However, these\\nmethods heavily rely on large, paired datasets, which is prohibitive for their\\nuse in scenarios where paired data does not exist, or there is only a small\\namount available. In contrast, image generation methods can work well on very\\nsmall datasets, and can find mappings between unpaired datasets, meaning an\\neffectively unlimited amount of paired synthetic data can be generated. In this\\nwork, we demonstrate that representation learning can be significantly improved\\nby synthetically generating paired information, both compared to training on\\neither single-modality (up to 4.4x error reduction) or authentic multi-modal\\npaired datasets (up to 5.6x error reduction).\",\"PeriodicalId\":501572,\"journal\":{\"name\":\"arXiv - QuanBio - Tissues and Organs\",\"volume\":\"91 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-03-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - QuanBio - Tissues and Organs\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2403.05220\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuanBio - Tissues and Organs","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2403.05220","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Synthetic Privileged Information Enhances Medical Image Representation Learning
Multimodal self-supervised representation learning has consistently proven to
be a highly effective method in medical image analysis, offering strong task
performance and producing biologically informed insights. However, these
methods heavily rely on large, paired datasets, which is prohibitive for their
use in scenarios where paired data does not exist, or there is only a small
amount available. In contrast, image generation methods can work well on very
small datasets, and can find mappings between unpaired datasets, meaning an
effectively unlimited amount of paired synthetic data can be generated. In this
work, we demonstrate that representation learning can be significantly improved
by synthetically generating paired information, both compared to training on
either single-modality (up to 4.4x error reduction) or authentic multi-modal
paired datasets (up to 5.6x error reduction).