H. K. Bhat, Aashish Mukund, S. Nagaraj, R. Prakash
{"title":"用于肺癌三维分割的u形变形器","authors":"H. K. Bhat, Aashish Mukund, S. Nagaraj, R. Prakash","doi":"10.1109/ICKECS56523.2022.10059627","DOIUrl":null,"url":null,"abstract":"3-Dimensional (3D) image segmentation in medical images is essential for early detection and diagnosis of diseases. It also aids in effective monitoring and treatment preparation. Traditional methods of delineating the image manually requires anatomical knowledge and is error-prone, cumbersome and expensive. Deep learning methods, especially V-shaped Convolutional Neural Network (CNN) architectures have achieved state-of-the-art performance on 2-Dimensional (2D) clinical image data. However, when it comes to 3D medical images, they suffer from anisotropy which is non-homogeneity in all directions. This paper shows that conventional convolution-based networks are insufficient to accurately segment this kind of data and proposes a ‘U-shaped’ transformer-based network, leveraging the self-attention mechanism to achieve better segmentation results. The proposed model outperforms baseline convolution-based models in 3D lung cancer segmentation.","PeriodicalId":171432,"journal":{"name":"2022 International Conference on Knowledge Engineering and Communication Systems (ICKES)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"U-shaped Transformers for 3D Lung Cancer Segmentation\",\"authors\":\"H. K. Bhat, Aashish Mukund, S. Nagaraj, R. Prakash\",\"doi\":\"10.1109/ICKECS56523.2022.10059627\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"3-Dimensional (3D) image segmentation in medical images is essential for early detection and diagnosis of diseases. It also aids in effective monitoring and treatment preparation. Traditional methods of delineating the image manually requires anatomical knowledge and is error-prone, cumbersome and expensive. Deep learning methods, especially V-shaped Convolutional Neural Network (CNN) architectures have achieved state-of-the-art performance on 2-Dimensional (2D) clinical image data. However, when it comes to 3D medical images, they suffer from anisotropy which is non-homogeneity in all directions. This paper shows that conventional convolution-based networks are insufficient to accurately segment this kind of data and proposes a ‘U-shaped’ transformer-based network, leveraging the self-attention mechanism to achieve better segmentation results. The proposed model outperforms baseline convolution-based models in 3D lung cancer segmentation.\",\"PeriodicalId\":171432,\"journal\":{\"name\":\"2022 International Conference on Knowledge Engineering and Communication Systems (ICKES)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-12-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 International Conference on Knowledge Engineering and Communication Systems (ICKES)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICKECS56523.2022.10059627\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Knowledge Engineering and Communication Systems (ICKES)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICKECS56523.2022.10059627","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
U-shaped Transformers for 3D Lung Cancer Segmentation
3-Dimensional (3D) image segmentation in medical images is essential for early detection and diagnosis of diseases. It also aids in effective monitoring and treatment preparation. Traditional methods of delineating the image manually requires anatomical knowledge and is error-prone, cumbersome and expensive. Deep learning methods, especially V-shaped Convolutional Neural Network (CNN) architectures have achieved state-of-the-art performance on 2-Dimensional (2D) clinical image data. However, when it comes to 3D medical images, they suffer from anisotropy which is non-homogeneity in all directions. This paper shows that conventional convolution-based networks are insufficient to accurately segment this kind of data and proposes a ‘U-shaped’ transformer-based network, leveraging the self-attention mechanism to achieve better segmentation results. The proposed model outperforms baseline convolution-based models in 3D lung cancer segmentation.