Thanh-Hai Tran, Phuong Thi Tuyet Nguyen, Duc-Huy Tran, X. Manh, Danh H. Vu, Nguyen-Khang Ho, Khanh-Linh Do, Van-Tuan Nguyen, Long-Thuy Nguyen, V. Dao, Hai Vu
{"title":"上消化道内镜图像解剖标志的分类","authors":"Thanh-Hai Tran, Phuong Thi Tuyet Nguyen, Duc-Huy Tran, X. Manh, Danh H. Vu, Nguyen-Khang Ho, Khanh-Linh Do, Van-Tuan Nguyen, Long-Thuy Nguyen, V. Dao, Hai Vu","doi":"10.1109/NICS54270.2021.9701513","DOIUrl":null,"url":null,"abstract":"In this paper, we propose a framework that automatically classifies anatomical landmarks of Upper GastroIntestinal Endoscopy (UGIE). This framework aims to select the best deep neural network in terms of both criteria of classification performances and computational costs. We investigate two lightweight deep neural networks that are ResNet-18, MobileNet-V2 to learn hidden discriminant features for multi classification task. In addition, because convolutional neural networks (CNNs) are data hungry, we examine various data augmentation (DA) techniques such as Brightness and Contrast Transformation (BaC), Geometric Transformation (GeoT), and Variational Auto-Encoder (VAE). Impacts of these DA schemes are evaluated for both CNN models. The experiments are conducted on a self collected dataset of 3700 endoscopic images which contains 10 anatomical landmarks of UGIE. The results show outstanding performances of both models thanks to DA techniques compared to the original data usage. The best sensitivity is 97.43% and specificity is 99.71% using MobileNet-V2 with Geometric Transformation based DA technique at a frame-rate of 21fps. These results highlight the best model which has significant potential for developing computer-aided esophagogastroduodenoscopy (EGD) diagnostic systems.","PeriodicalId":296963,"journal":{"name":"2021 8th NAFOSTED Conference on Information and Computer Science (NICS)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Classification of anatomical landmarks from upper gastrointestinal endoscopic images⋆\",\"authors\":\"Thanh-Hai Tran, Phuong Thi Tuyet Nguyen, Duc-Huy Tran, X. Manh, Danh H. Vu, Nguyen-Khang Ho, Khanh-Linh Do, Van-Tuan Nguyen, Long-Thuy Nguyen, V. Dao, Hai Vu\",\"doi\":\"10.1109/NICS54270.2021.9701513\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we propose a framework that automatically classifies anatomical landmarks of Upper GastroIntestinal Endoscopy (UGIE). This framework aims to select the best deep neural network in terms of both criteria of classification performances and computational costs. We investigate two lightweight deep neural networks that are ResNet-18, MobileNet-V2 to learn hidden discriminant features for multi classification task. In addition, because convolutional neural networks (CNNs) are data hungry, we examine various data augmentation (DA) techniques such as Brightness and Contrast Transformation (BaC), Geometric Transformation (GeoT), and Variational Auto-Encoder (VAE). Impacts of these DA schemes are evaluated for both CNN models. The experiments are conducted on a self collected dataset of 3700 endoscopic images which contains 10 anatomical landmarks of UGIE. The results show outstanding performances of both models thanks to DA techniques compared to the original data usage. The best sensitivity is 97.43% and specificity is 99.71% using MobileNet-V2 with Geometric Transformation based DA technique at a frame-rate of 21fps. These results highlight the best model which has significant potential for developing computer-aided esophagogastroduodenoscopy (EGD) diagnostic systems.\",\"PeriodicalId\":296963,\"journal\":{\"name\":\"2021 8th NAFOSTED Conference on Information and Computer Science (NICS)\",\"volume\":\"15 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 8th NAFOSTED Conference on Information and Computer Science (NICS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/NICS54270.2021.9701513\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 8th NAFOSTED Conference on Information and Computer Science (NICS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NICS54270.2021.9701513","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Classification of anatomical landmarks from upper gastrointestinal endoscopic images⋆
In this paper, we propose a framework that automatically classifies anatomical landmarks of Upper GastroIntestinal Endoscopy (UGIE). This framework aims to select the best deep neural network in terms of both criteria of classification performances and computational costs. We investigate two lightweight deep neural networks that are ResNet-18, MobileNet-V2 to learn hidden discriminant features for multi classification task. In addition, because convolutional neural networks (CNNs) are data hungry, we examine various data augmentation (DA) techniques such as Brightness and Contrast Transformation (BaC), Geometric Transformation (GeoT), and Variational Auto-Encoder (VAE). Impacts of these DA schemes are evaluated for both CNN models. The experiments are conducted on a self collected dataset of 3700 endoscopic images which contains 10 anatomical landmarks of UGIE. The results show outstanding performances of both models thanks to DA techniques compared to the original data usage. The best sensitivity is 97.43% and specificity is 99.71% using MobileNet-V2 with Geometric Transformation based DA technique at a frame-rate of 21fps. These results highlight the best model which has significant potential for developing computer-aided esophagogastroduodenoscopy (EGD) diagnostic systems.