Shota Suzuki, Takafumi Katayama, Tian Song, T. Shimamoto
{"title":"基于域自适应的交叉视频语义分割","authors":"Shota Suzuki, Takafumi Katayama, Tian Song, T. Shimamoto","doi":"10.1109/ITC-CSCC58803.2023.10212842","DOIUrl":null,"url":null,"abstract":"In recent years, semantic segmentation, which is one of the image recognition technologies for automatic driving, has attracted attention. Semantic segmentation can perform high accuracy object detection by discriminating classes with pixel level precision. However, training data for segmentation usually requires extensive manual labor because a corresponding label has to be assigned to each pixel. Currently, using a computer graphics (CG) dataset makes it easier to create supervised data. However, when inference images are real-world, intersection over union (IoU) is greatly reduced by the variation of the domain. In this work, an unsupervised domain adaptation (UDA) training method is proposed to achieve efficient intersection video segmentation models for each target domain. The inference is performed for intersection images with relatively complex situations. The simulation results show that the proposed algorithm for semantic segmentation of intersection videos achieve IoU scores comparable to supervised learning for classes that are prominently displayed on the screen.","PeriodicalId":220939,"journal":{"name":"2023 International Technical Conference on Circuits/Systems, Computers, and Communications (ITC-CSCC)","volume":"361 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Video Semantic Segmentation for Intersection by Domain Adaptation\",\"authors\":\"Shota Suzuki, Takafumi Katayama, Tian Song, T. Shimamoto\",\"doi\":\"10.1109/ITC-CSCC58803.2023.10212842\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In recent years, semantic segmentation, which is one of the image recognition technologies for automatic driving, has attracted attention. Semantic segmentation can perform high accuracy object detection by discriminating classes with pixel level precision. However, training data for segmentation usually requires extensive manual labor because a corresponding label has to be assigned to each pixel. Currently, using a computer graphics (CG) dataset makes it easier to create supervised data. However, when inference images are real-world, intersection over union (IoU) is greatly reduced by the variation of the domain. In this work, an unsupervised domain adaptation (UDA) training method is proposed to achieve efficient intersection video segmentation models for each target domain. The inference is performed for intersection images with relatively complex situations. The simulation results show that the proposed algorithm for semantic segmentation of intersection videos achieve IoU scores comparable to supervised learning for classes that are prominently displayed on the screen.\",\"PeriodicalId\":220939,\"journal\":{\"name\":\"2023 International Technical Conference on Circuits/Systems, Computers, and Communications (ITC-CSCC)\",\"volume\":\"361 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-06-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 International Technical Conference on Circuits/Systems, Computers, and Communications (ITC-CSCC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ITC-CSCC58803.2023.10212842\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Technical Conference on Circuits/Systems, Computers, and Communications (ITC-CSCC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITC-CSCC58803.2023.10212842","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Video Semantic Segmentation for Intersection by Domain Adaptation
In recent years, semantic segmentation, which is one of the image recognition technologies for automatic driving, has attracted attention. Semantic segmentation can perform high accuracy object detection by discriminating classes with pixel level precision. However, training data for segmentation usually requires extensive manual labor because a corresponding label has to be assigned to each pixel. Currently, using a computer graphics (CG) dataset makes it easier to create supervised data. However, when inference images are real-world, intersection over union (IoU) is greatly reduced by the variation of the domain. In this work, an unsupervised domain adaptation (UDA) training method is proposed to achieve efficient intersection video segmentation models for each target domain. The inference is performed for intersection images with relatively complex situations. The simulation results show that the proposed algorithm for semantic segmentation of intersection videos achieve IoU scores comparable to supervised learning for classes that are prominently displayed on the screen.