Vanshali Sharma, Pradipta Sasmal, M. Bhuyan, P. Das
{"title":"Keyframe Selection from Colonoscopy Videos to Enhance Visualization for Polyp Detection","authors":"Vanshali Sharma, Pradipta Sasmal, M. Bhuyan, P. Das","doi":"10.1109/IV56949.2022.00076","DOIUrl":null,"url":null,"abstract":"Colonoscopy video acquisition and recording have been increasingly performed for comprehensive diagnosis and retrospective analysis of colorectal cancer (CRC). Reviewing video streams helps detect and inspect polyps, the precursor to CRC. However, visualizing these streams in their raw form puts a considerable burden on clinicians as most of the frames are clinically insignificant and are not useful for pathological interpretation. For improved visualization of diagnostically significant information, we have proposed an automated framework that discards the uninformative frames from raw videos. Our approach initially extracts high-quality colonoscopy frames using a deep learning model to assist clinicians in visualizing data in a refined form. Subsequently, our work validates the effectiveness of keyframe selection by employing polyp detection models. All the evaluations are performed either patient-wise or cross-dataset to suffice the real-time requirements. Experimental results show that the keyframe extraction saves reviewing time and enhances the detection performances. The proposed approach achieves a polyp detection F1-score of 79.78% (patient-wise) and 89.22% (cross-dataset) on the SUN and CVC-VideoClinicDB databases, respectively.","PeriodicalId":153161,"journal":{"name":"2022 26th International Conference Information Visualisation (IV)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 26th International Conference Information Visualisation (IV)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IV56949.2022.00076","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Colonoscopy video acquisition and recording have been increasingly performed for comprehensive diagnosis and retrospective analysis of colorectal cancer (CRC). Reviewing video streams helps detect and inspect polyps, the precursor to CRC. However, visualizing these streams in their raw form puts a considerable burden on clinicians as most of the frames are clinically insignificant and are not useful for pathological interpretation. For improved visualization of diagnostically significant information, we have proposed an automated framework that discards the uninformative frames from raw videos. Our approach initially extracts high-quality colonoscopy frames using a deep learning model to assist clinicians in visualizing data in a refined form. Subsequently, our work validates the effectiveness of keyframe selection by employing polyp detection models. All the evaluations are performed either patient-wise or cross-dataset to suffice the real-time requirements. Experimental results show that the keyframe extraction saves reviewing time and enhances the detection performances. The proposed approach achieves a polyp detection F1-score of 79.78% (patient-wise) and 89.22% (cross-dataset) on the SUN and CVC-VideoClinicDB databases, respectively.