Yu Zhao, Xiaoyan Shen, Jiadong Chen, Wei Qian, He Ma, Liang Sang
{"title":"用于低对比度医学图像分割的 δARD 损失","authors":"Yu Zhao, Xiaoyan Shen, Jiadong Chen, Wei Qian, He Ma, Liang Sang","doi":"10.1088/2632-2153/ad1d06","DOIUrl":null,"url":null,"abstract":"\n Purpose Medical image segmentation is essential to image-based disease analysis and has proven to be significantly helpful for doctors to make decisions. Due to the low-contrast of some medical images, the accurate segmentation of medical images has always been a challenging problem. The experiment found that UNet with current loss functions cannot capture subtle information in target contours or regions in low-contrast medical images, which are crucial for subsequent disease diagnosis. Methods We propose a robust loss by incorporating the difference in average radial derivative (ARD), length and region area to further help the network to achieve more accurate segmentation results. We evaluated the proposed loss function using UNet as the base segmentation network compared to five conventional loss functions on one private and four public medical image datasets. Results Experimental results illustrate that UNet with the proposed loss function can achieve the best segmentation performance, even better than the outstanding deep learning models with original loss functions. Furthermore, three representative datasets were chosen to validate the effectiveness of the proposed δARD loss function with seven different models. Conclusion The experiments revealed δARD loss's plug-and-play feature and its robustness over multiple models and datasets.","PeriodicalId":503691,"journal":{"name":"Machine Learning: Science and Technology","volume":"4 12","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"δARD loss for low-contrast medical image segmentation\",\"authors\":\"Yu Zhao, Xiaoyan Shen, Jiadong Chen, Wei Qian, He Ma, Liang Sang\",\"doi\":\"10.1088/2632-2153/ad1d06\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"\\n Purpose Medical image segmentation is essential to image-based disease analysis and has proven to be significantly helpful for doctors to make decisions. Due to the low-contrast of some medical images, the accurate segmentation of medical images has always been a challenging problem. The experiment found that UNet with current loss functions cannot capture subtle information in target contours or regions in low-contrast medical images, which are crucial for subsequent disease diagnosis. Methods We propose a robust loss by incorporating the difference in average radial derivative (ARD), length and region area to further help the network to achieve more accurate segmentation results. We evaluated the proposed loss function using UNet as the base segmentation network compared to five conventional loss functions on one private and four public medical image datasets. Results Experimental results illustrate that UNet with the proposed loss function can achieve the best segmentation performance, even better than the outstanding deep learning models with original loss functions. Furthermore, three representative datasets were chosen to validate the effectiveness of the proposed δARD loss function with seven different models. Conclusion The experiments revealed δARD loss's plug-and-play feature and its robustness over multiple models and datasets.\",\"PeriodicalId\":503691,\"journal\":{\"name\":\"Machine Learning: Science and Technology\",\"volume\":\"4 12\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-01-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Machine Learning: Science and Technology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1088/2632-2153/ad1d06\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Machine Learning: Science and Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1088/2632-2153/ad1d06","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
δARD loss for low-contrast medical image segmentation
Purpose Medical image segmentation is essential to image-based disease analysis and has proven to be significantly helpful for doctors to make decisions. Due to the low-contrast of some medical images, the accurate segmentation of medical images has always been a challenging problem. The experiment found that UNet with current loss functions cannot capture subtle information in target contours or regions in low-contrast medical images, which are crucial for subsequent disease diagnosis. Methods We propose a robust loss by incorporating the difference in average radial derivative (ARD), length and region area to further help the network to achieve more accurate segmentation results. We evaluated the proposed loss function using UNet as the base segmentation network compared to five conventional loss functions on one private and four public medical image datasets. Results Experimental results illustrate that UNet with the proposed loss function can achieve the best segmentation performance, even better than the outstanding deep learning models with original loss functions. Furthermore, three representative datasets were chosen to validate the effectiveness of the proposed δARD loss function with seven different models. Conclusion The experiments revealed δARD loss's plug-and-play feature and its robustness over multiple models and datasets.