[A medical visual question answering approach based on co-attention networks].

Q4 Medicine
Wencheng Cui, Wentao Shi, Hong Shao
{"title":"[A medical visual question answering approach based on co-attention networks].","authors":"Wencheng Cui, Wentao Shi, Hong Shao","doi":"10.7507/1001-5515.202307057","DOIUrl":null,"url":null,"abstract":"<p><p>Recent studies have introduced attention models for medical visual question answering (MVQA). In medical research, not only is the modeling of \"visual attention\" crucial, but the modeling of \"question attention\" is equally significant. To facilitate bidirectional reasoning in the attention processes involving medical images and questions, a new MVQA architecture, named MCAN, has been proposed. This architecture incorporated a cross-modal co-attention network, FCAF, which identifies key words in questions and principal parts in images. Through a meta-learning channel attention module (MLCA), weights were adaptively assigned to each word and region, reflecting the model's focus on specific words and regions during reasoning. Additionally, this study specially designed and developed a medical domain-specific word embedding model, Med-GloVe, to further enhance the model's accuracy and practical value. Experimental results indicated that MCAN proposed in this study improved the accuracy by 7.7% on free-form questions in the Path-VQA dataset, and by 4.4% on closed-form questions in the VQA-RAD dataset, which effectively improves the accuracy of the medical vision question answer.</p>","PeriodicalId":39324,"journal":{"name":"生物医学工程学杂志","volume":"41 3","pages":"560-568"},"PeriodicalIF":0.0000,"publicationDate":"2024-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11208638/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"生物医学工程学杂志","FirstCategoryId":"1087","ListUrlMain":"https://doi.org/10.7507/1001-5515.202307057","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"Medicine","Score":null,"Total":0}
引用次数: 0

Abstract

Recent studies have introduced attention models for medical visual question answering (MVQA). In medical research, not only is the modeling of "visual attention" crucial, but the modeling of "question attention" is equally significant. To facilitate bidirectional reasoning in the attention processes involving medical images and questions, a new MVQA architecture, named MCAN, has been proposed. This architecture incorporated a cross-modal co-attention network, FCAF, which identifies key words in questions and principal parts in images. Through a meta-learning channel attention module (MLCA), weights were adaptively assigned to each word and region, reflecting the model's focus on specific words and regions during reasoning. Additionally, this study specially designed and developed a medical domain-specific word embedding model, Med-GloVe, to further enhance the model's accuracy and practical value. Experimental results indicated that MCAN proposed in this study improved the accuracy by 7.7% on free-form questions in the Path-VQA dataset, and by 4.4% on closed-form questions in the VQA-RAD dataset, which effectively improves the accuracy of the medical vision question answer.

[基于共注意力网络的医学视觉问题解答方法]。
最近的研究为医学视觉问题解答(MVQA)引入了注意力模型。在医学研究中,不仅 "视觉注意力 "的建模至关重要,"问题注意力 "的建模也同样重要。为了促进涉及医学图像和问题的注意过程中的双向推理,我们提出了一种名为 MCAN 的新型 MVQA 架构。该架构包含一个跨模态协同注意网络 FCAF,可识别问题中的关键词和图像中的主要部分。通过元学习通道注意模块(MLCA),自适应地为每个单词和区域分配权重,以反映模型在推理过程中对特定单词和区域的关注。此外,本研究还专门设计和开发了医学领域专用的单词嵌入模型 Med-GloVe,以进一步提高模型的准确性和实用价值。实验结果表明,本研究提出的 MCAN 在 Path-VQA 数据集的自由形式问题上提高了 7.7% 的准确率,在 VQA-RAD 数据集的封闭形式问题上提高了 4.4% 的准确率,有效地提高了医学视觉问题答案的准确率。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
生物医学工程学杂志
生物医学工程学杂志 Medicine-Medicine (all)
CiteScore
0.80
自引率
0.00%
发文量
4868
期刊介绍:
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信