{"title":"利用卷积神经网络进行水下三维物体重构以估算鱼体长度","authors":"Naomi Ubiña, Sin-Yi Cai, Shyi-Chyi Cheng, Chin-Chun Chang, Yi-Zeng Hsieh","doi":"10.1109/ISPACS51563.2021.9651057","DOIUrl":null,"url":null,"abstract":"The challenge in stereo matching for underwater 3D object modeling is to compute dense depth data with the minimal smoothness at high frame rate. To address this issue, in this paper we propose an object based stereo matching for underwater 3D fish reconstruction using convolutional neural networks (CNNs). For each image in a stereo frame, an instance segmentation CNN is used to segment fish objects from the background. The set of fish objects in the left image is matched against those in the right to detect the object pairs using the proposed support-weights approach. For each pair, the common disparity value is then computed. Next, fish objects in these images are cropped and matched to do the pixel-wise residual disparity computation using the video interpolation CNN. The computed fish disparities and depth values are finally used to estimate the sizes of fish. Instead of estimating the fish length using a single frame, we track each fish across frames of the input stereo video to compute the fish length frame by frame. The mean fish length is finally computed as the result. An underwater dataset with the fish actual length measured by human is constructed to verify the effectiveness of our approach. Experimental results show that the error rate of the proposed approach is less than 6%.","PeriodicalId":359822,"journal":{"name":"2021 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Underwater 3D Object Reconstruction for Fish Length Estimation Using Convolutional Neural Networks\",\"authors\":\"Naomi Ubiña, Sin-Yi Cai, Shyi-Chyi Cheng, Chin-Chun Chang, Yi-Zeng Hsieh\",\"doi\":\"10.1109/ISPACS51563.2021.9651057\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The challenge in stereo matching for underwater 3D object modeling is to compute dense depth data with the minimal smoothness at high frame rate. To address this issue, in this paper we propose an object based stereo matching for underwater 3D fish reconstruction using convolutional neural networks (CNNs). For each image in a stereo frame, an instance segmentation CNN is used to segment fish objects from the background. The set of fish objects in the left image is matched against those in the right to detect the object pairs using the proposed support-weights approach. For each pair, the common disparity value is then computed. Next, fish objects in these images are cropped and matched to do the pixel-wise residual disparity computation using the video interpolation CNN. The computed fish disparities and depth values are finally used to estimate the sizes of fish. Instead of estimating the fish length using a single frame, we track each fish across frames of the input stereo video to compute the fish length frame by frame. The mean fish length is finally computed as the result. An underwater dataset with the fish actual length measured by human is constructed to verify the effectiveness of our approach. Experimental results show that the error rate of the proposed approach is less than 6%.\",\"PeriodicalId\":359822,\"journal\":{\"name\":\"2021 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-11-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISPACS51563.2021.9651057\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISPACS51563.2021.9651057","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Underwater 3D Object Reconstruction for Fish Length Estimation Using Convolutional Neural Networks
The challenge in stereo matching for underwater 3D object modeling is to compute dense depth data with the minimal smoothness at high frame rate. To address this issue, in this paper we propose an object based stereo matching for underwater 3D fish reconstruction using convolutional neural networks (CNNs). For each image in a stereo frame, an instance segmentation CNN is used to segment fish objects from the background. The set of fish objects in the left image is matched against those in the right to detect the object pairs using the proposed support-weights approach. For each pair, the common disparity value is then computed. Next, fish objects in these images are cropped and matched to do the pixel-wise residual disparity computation using the video interpolation CNN. The computed fish disparities and depth values are finally used to estimate the sizes of fish. Instead of estimating the fish length using a single frame, we track each fish across frames of the input stereo video to compute the fish length frame by frame. The mean fish length is finally computed as the result. An underwater dataset with the fish actual length measured by human is constructed to verify the effectiveness of our approach. Experimental results show that the error rate of the proposed approach is less than 6%.