{"title":"一种新的扩展卷积神经网络道路场景分割模型","authors":"Yachao Zhang, Yuxia Yuan","doi":"10.4108/eai.27-1-2022.173164","DOIUrl":null,"url":null,"abstract":"Road scene understanding is one of the important modules in the field of autonomous driving. It can provide more information about roads and play an important role in building high-precision maps and real-time planning. Among them, semantic segmentation can assign category information to each pixel of image, which is the most commonly used method in automatic driving scene understanding. However, most commonly used semantic segmentation algorithms cannot achieve a good balance between speed and precision. In this paper, a road scene segmentation model based on dilated convolutional neural network is constructed. The model consists of a front-end module and a context module. The front-end module is an improved structure of VGG-16 fused dilated convolution, and the context module is a cascade of dilated convolution layers with different expansion coefficients, which is trained by a two-stage training method. The network proposed in this paper can run in real time and ensure the accuracy to meet the requirements of practical applications, and has been verified and analyzed on Cityscapes data set.","PeriodicalId":1,"journal":{"name":"Accounts of Chemical Research","volume":null,"pages":null},"PeriodicalIF":16.4000,"publicationDate":"2022-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"A novel dilated convolutional neural network model for road scene segmentation\",\"authors\":\"Yachao Zhang, Yuxia Yuan\",\"doi\":\"10.4108/eai.27-1-2022.173164\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Road scene understanding is one of the important modules in the field of autonomous driving. It can provide more information about roads and play an important role in building high-precision maps and real-time planning. Among them, semantic segmentation can assign category information to each pixel of image, which is the most commonly used method in automatic driving scene understanding. However, most commonly used semantic segmentation algorithms cannot achieve a good balance between speed and precision. In this paper, a road scene segmentation model based on dilated convolutional neural network is constructed. The model consists of a front-end module and a context module. The front-end module is an improved structure of VGG-16 fused dilated convolution, and the context module is a cascade of dilated convolution layers with different expansion coefficients, which is trained by a two-stage training method. The network proposed in this paper can run in real time and ensure the accuracy to meet the requirements of practical applications, and has been verified and analyzed on Cityscapes data set.\",\"PeriodicalId\":1,\"journal\":{\"name\":\"Accounts of Chemical Research\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":16.4000,\"publicationDate\":\"2022-01-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Accounts of Chemical Research\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.4108/eai.27-1-2022.173164\",\"RegionNum\":1,\"RegionCategory\":\"化学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"CHEMISTRY, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Accounts of Chemical Research","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.4108/eai.27-1-2022.173164","RegionNum":1,"RegionCategory":"化学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"CHEMISTRY, MULTIDISCIPLINARY","Score":null,"Total":0}
A novel dilated convolutional neural network model for road scene segmentation
Road scene understanding is one of the important modules in the field of autonomous driving. It can provide more information about roads and play an important role in building high-precision maps and real-time planning. Among them, semantic segmentation can assign category information to each pixel of image, which is the most commonly used method in automatic driving scene understanding. However, most commonly used semantic segmentation algorithms cannot achieve a good balance between speed and precision. In this paper, a road scene segmentation model based on dilated convolutional neural network is constructed. The model consists of a front-end module and a context module. The front-end module is an improved structure of VGG-16 fused dilated convolution, and the context module is a cascade of dilated convolution layers with different expansion coefficients, which is trained by a two-stage training method. The network proposed in this paper can run in real time and ensure the accuracy to meet the requirements of practical applications, and has been verified and analyzed on Cityscapes data set.
期刊介绍:
Accounts of Chemical Research presents short, concise and critical articles offering easy-to-read overviews of basic research and applications in all areas of chemistry and biochemistry. These short reviews focus on research from the author’s own laboratory and are designed to teach the reader about a research project. In addition, Accounts of Chemical Research publishes commentaries that give an informed opinion on a current research problem. Special Issues online are devoted to a single topic of unusual activity and significance.
Accounts of Chemical Research replaces the traditional article abstract with an article "Conspectus." These entries synopsize the research affording the reader a closer look at the content and significance of an article. Through this provision of a more detailed description of the article contents, the Conspectus enhances the article's discoverability by search engines and the exposure for the research.