Chuanbo Zhou, Guoan Yang, Zhengzhi Lu, Deyang Liu, Yong Yang
{"title":"A Noise-robust Feature Fusion Model Combining Non-local Attention for Material Recognition","authors":"Chuanbo Zhou, Guoan Yang, Zhengzhi Lu, Deyang Liu, Yong Yang","doi":"10.1145/3512388.3512450","DOIUrl":null,"url":null,"abstract":"Material recognition, as an important task of computer vision, is hugely challenging, due to large intra-class variances and small inter-class variances between material images. To address those recognition problems, multi-scale feature fusion methods based on deep convolutional neural networks are presented, which has been widely studied in recent years. However, the past research works paid too much attention to the local features of the image, while ignoring the non-local features that are also crucial for fine image recognition tasks such as material recognition. In this paper, Non-local Attentional Feature Fusion Network (NLA-FFNet) is proposed that combines local and non-local feature of images to improve the feature representation capability. Firstly, we utilize the pre-trained deep convolutional neural network to extract the image feature. Secondly, a Multilayer Non-local Attention (MNLA) block is designed to generate a non-local attention map which represents the long-range dependencies between features of different positions. Therefore, it can achieve stronger noise-robustness of model and better ability to represent fine features. Finally, combined our Multilayer Non-local Attention block with bilinear pooling which has been proved to be effective for feature fusion, we propose a deep neural network framework, NLA-FFNet, with noise-robust multi-layer feature fusion. Experiment prove that our model can achieve a competitive classification accuracy in material image recognition, and has stronger noise-robustness at the same time.","PeriodicalId":434878,"journal":{"name":"Proceedings of the 2022 5th International Conference on Image and Graphics Processing","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2022 5th International Conference on Image and Graphics Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3512388.3512450","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
Material recognition, as an important task of computer vision, is hugely challenging, due to large intra-class variances and small inter-class variances between material images. To address those recognition problems, multi-scale feature fusion methods based on deep convolutional neural networks are presented, which has been widely studied in recent years. However, the past research works paid too much attention to the local features of the image, while ignoring the non-local features that are also crucial for fine image recognition tasks such as material recognition. In this paper, Non-local Attentional Feature Fusion Network (NLA-FFNet) is proposed that combines local and non-local feature of images to improve the feature representation capability. Firstly, we utilize the pre-trained deep convolutional neural network to extract the image feature. Secondly, a Multilayer Non-local Attention (MNLA) block is designed to generate a non-local attention map which represents the long-range dependencies between features of different positions. Therefore, it can achieve stronger noise-robustness of model and better ability to represent fine features. Finally, combined our Multilayer Non-local Attention block with bilinear pooling which has been proved to be effective for feature fusion, we propose a deep neural network framework, NLA-FFNet, with noise-robust multi-layer feature fusion. Experiment prove that our model can achieve a competitive classification accuracy in material image recognition, and has stronger noise-robustness at the same time.