{"title":"用于高级视觉字幕的属性细化注意力融合网络","authors":"Md. Bipul Hossen , Zhongfu Ye , Md. Shamim Hossain , Md. Imran Hossain","doi":"10.1016/j.dsp.2025.105155","DOIUrl":null,"url":null,"abstract":"<div><div>Visual captioning, at the nexus of computer vision and natural language processing, is one of the pivotal aspects of multimedia content understanding, demands precise and contextually fitting image descriptions. Attribute-based approaches with attention mechanisms are effective in this realm. However, many of these approaches struggle to capture semantic details due to the prediction of irrelevant attributes and reduced performance. In response to these challenges, we propose an innovative solution: the Attribute Refinement Attention Fusion Network (ARAFNet), which aims to produce significant captions by distinctly identifying major objects and background information. The model features a comprehensive Attribute Refinement Attention (ARA) module, equipped with an attribute attention mechanism, which interactively extracts the most important attributes according to the linguistic context. Diverse attributes are employed at different time steps, enhancing the model's capability to utilize semantic features effectively while also filtering out irrelevant attribute words, thereby enhancing the precision of semantic guidance. An integrated fusion mechanism is then introduced to narrow the semantic gap between visual and attribute features. Finally, this fusion mechanism combined with the language LSTM to generate precise and contextually relevant captions. Extensive experimentation demonstrates our model's superiority over advanced counterparts, achieving an average CIDEr-D score of 11.88% on the Flickr30K dataset and 11.25% on the MS-COCO dataset through cross-entropy optimization. The ARAFNet model consistently outperforms the baseline model across a diverse range of evaluation metrics and makes a significant contribution to the field of image captioning precision. The implementing code and associated materials will be published at <span><span>https://github.com/mdbipu/ARAFNet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"162 ","pages":"Article 105155"},"PeriodicalIF":2.9000,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ARAFNet: An attribute refinement attention fusion network for advanced visual captioning\",\"authors\":\"Md. Bipul Hossen , Zhongfu Ye , Md. Shamim Hossain , Md. Imran Hossain\",\"doi\":\"10.1016/j.dsp.2025.105155\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Visual captioning, at the nexus of computer vision and natural language processing, is one of the pivotal aspects of multimedia content understanding, demands precise and contextually fitting image descriptions. Attribute-based approaches with attention mechanisms are effective in this realm. However, many of these approaches struggle to capture semantic details due to the prediction of irrelevant attributes and reduced performance. In response to these challenges, we propose an innovative solution: the Attribute Refinement Attention Fusion Network (ARAFNet), which aims to produce significant captions by distinctly identifying major objects and background information. The model features a comprehensive Attribute Refinement Attention (ARA) module, equipped with an attribute attention mechanism, which interactively extracts the most important attributes according to the linguistic context. Diverse attributes are employed at different time steps, enhancing the model's capability to utilize semantic features effectively while also filtering out irrelevant attribute words, thereby enhancing the precision of semantic guidance. An integrated fusion mechanism is then introduced to narrow the semantic gap between visual and attribute features. Finally, this fusion mechanism combined with the language LSTM to generate precise and contextually relevant captions. Extensive experimentation demonstrates our model's superiority over advanced counterparts, achieving an average CIDEr-D score of 11.88% on the Flickr30K dataset and 11.25% on the MS-COCO dataset through cross-entropy optimization. The ARAFNet model consistently outperforms the baseline model across a diverse range of evaluation metrics and makes a significant contribution to the field of image captioning precision. The implementing code and associated materials will be published at <span><span>https://github.com/mdbipu/ARAFNet</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":51011,\"journal\":{\"name\":\"Digital Signal Processing\",\"volume\":\"162 \",\"pages\":\"Article 105155\"},\"PeriodicalIF\":2.9000,\"publicationDate\":\"2025-03-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Digital Signal Processing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1051200425001770\",\"RegionNum\":3,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Digital Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1051200425001770","RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
ARAFNet: An attribute refinement attention fusion network for advanced visual captioning
Visual captioning, at the nexus of computer vision and natural language processing, is one of the pivotal aspects of multimedia content understanding, demands precise and contextually fitting image descriptions. Attribute-based approaches with attention mechanisms are effective in this realm. However, many of these approaches struggle to capture semantic details due to the prediction of irrelevant attributes and reduced performance. In response to these challenges, we propose an innovative solution: the Attribute Refinement Attention Fusion Network (ARAFNet), which aims to produce significant captions by distinctly identifying major objects and background information. The model features a comprehensive Attribute Refinement Attention (ARA) module, equipped with an attribute attention mechanism, which interactively extracts the most important attributes according to the linguistic context. Diverse attributes are employed at different time steps, enhancing the model's capability to utilize semantic features effectively while also filtering out irrelevant attribute words, thereby enhancing the precision of semantic guidance. An integrated fusion mechanism is then introduced to narrow the semantic gap between visual and attribute features. Finally, this fusion mechanism combined with the language LSTM to generate precise and contextually relevant captions. Extensive experimentation demonstrates our model's superiority over advanced counterparts, achieving an average CIDEr-D score of 11.88% on the Flickr30K dataset and 11.25% on the MS-COCO dataset through cross-entropy optimization. The ARAFNet model consistently outperforms the baseline model across a diverse range of evaluation metrics and makes a significant contribution to the field of image captioning precision. The implementing code and associated materials will be published at https://github.com/mdbipu/ARAFNet.
期刊介绍:
Digital Signal Processing: A Review Journal is one of the oldest and most established journals in the field of signal processing yet it aims to be the most innovative. The Journal invites top quality research articles at the frontiers of research in all aspects of signal processing. Our objective is to provide a platform for the publication of ground-breaking research in signal processing with both academic and industrial appeal.
The journal has a special emphasis on statistical signal processing methodology such as Bayesian signal processing, and encourages articles on emerging applications of signal processing such as:
• big data• machine learning• internet of things• information security• systems biology and computational biology,• financial time series analysis,• autonomous vehicles,• quantum computing,• neuromorphic engineering,• human-computer interaction and intelligent user interfaces,• environmental signal processing,• geophysical signal processing including seismic signal processing,• chemioinformatics and bioinformatics,• audio, visual and performance arts,• disaster management and prevention,• renewable energy,