{"title":"基于语义分割比较框架的多尺度模型改进","authors":"Ting-Chen Hsu, Bor-Shen Lin","doi":"10.1109/taai54685.2021.00065","DOIUrl":null,"url":null,"abstract":"State-of-the-art models of semantic segmentation, based on global convolutional network, ASPP, self-attention, and so on, focus on capturing context information through the fusion of multi-scale features and integrating local features with large kernels. However, these models have not been compared yet in parallel to interpret their relative efficacies. This makes it difficult to further combine or improve these models due to their complicated network structures. In this paper, a general multi-scale framework of semantic image segmentation was proposed to investigate and compare the network structures of the models in parallel. Three alternative modules were proposed to improve these methods, and the experiments show the proposed modules can give superior segmentation results and achieve outstanding performance on Pascal VOC2012 images segmentation datasets. Additionally, this framework was shown to be flexible for integrating the multi-scale features and operations from different levels. Experimental results show that the low-level operation can extract local details and the high-level operation the overall contour, so the output features from different levels complement each other to improve the performance effectively.","PeriodicalId":343821,"journal":{"name":"2021 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)","volume":"73-74 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Improving Multi-Scale Models with A Comparative Framework for Semantic Segmentation\",\"authors\":\"Ting-Chen Hsu, Bor-Shen Lin\",\"doi\":\"10.1109/taai54685.2021.00065\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"State-of-the-art models of semantic segmentation, based on global convolutional network, ASPP, self-attention, and so on, focus on capturing context information through the fusion of multi-scale features and integrating local features with large kernels. However, these models have not been compared yet in parallel to interpret their relative efficacies. This makes it difficult to further combine or improve these models due to their complicated network structures. In this paper, a general multi-scale framework of semantic image segmentation was proposed to investigate and compare the network structures of the models in parallel. Three alternative modules were proposed to improve these methods, and the experiments show the proposed modules can give superior segmentation results and achieve outstanding performance on Pascal VOC2012 images segmentation datasets. Additionally, this framework was shown to be flexible for integrating the multi-scale features and operations from different levels. Experimental results show that the low-level operation can extract local details and the high-level operation the overall contour, so the output features from different levels complement each other to improve the performance effectively.\",\"PeriodicalId\":343821,\"journal\":{\"name\":\"2021 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)\",\"volume\":\"73-74 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/taai54685.2021.00065\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/taai54685.2021.00065","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Improving Multi-Scale Models with A Comparative Framework for Semantic Segmentation
State-of-the-art models of semantic segmentation, based on global convolutional network, ASPP, self-attention, and so on, focus on capturing context information through the fusion of multi-scale features and integrating local features with large kernels. However, these models have not been compared yet in parallel to interpret their relative efficacies. This makes it difficult to further combine or improve these models due to their complicated network structures. In this paper, a general multi-scale framework of semantic image segmentation was proposed to investigate and compare the network structures of the models in parallel. Three alternative modules were proposed to improve these methods, and the experiments show the proposed modules can give superior segmentation results and achieve outstanding performance on Pascal VOC2012 images segmentation datasets. Additionally, this framework was shown to be flexible for integrating the multi-scale features and operations from different levels. Experimental results show that the low-level operation can extract local details and the high-level operation the overall contour, so the output features from different levels complement each other to improve the performance effectively.