CollabAS2:使用基于变换器的协作模型加强阿拉伯语答案句子选择

IF 2.9 4区 综合性期刊 Q1 Multidisciplinary
Asma Aouichat, Ahmed Guessoum
{"title":"CollabAS2:使用基于变换器的协作模型加强阿拉伯语答案句子选择","authors":"Asma Aouichat, Ahmed Guessoum","doi":"10.1007/s13369-024-09345-3","DOIUrl":null,"url":null,"abstract":"<p>Accurately identifying pertinent text segments as answers to questions is crucial for optimizing question-answering systems, underscoring the pivotal role of precision in Answer Sentence Selection (AS2) modules. This study introduces an innovative AS2 module design leveraging the AraBERT transformer to encode inputs-one for the question and one for the candidate answer-with the goal of enhancing comprehension of both inputs. Each encoded input is subsequently processed in parallel by a collaborative layer employing two distinct deep learning models: a bidirectional long short-term memory (BiLSTM) and a convolutional neural network (CNN). This collaborative approach forms the AraBERT.Collab-BiLSTM/CNN model. Additionally, extensions to the study include AraBERT.Collab-BiLSTM/AVG, incorporating a BiLSTM and AVG collaboration layer, as well as the use of the AraELECTRA pre-trained model, yielding the AraELECTRA.Collab-BiLSTM/CNN and AraELECTRA.Collab-BiLSTM/AVG configurations. Furthermore, the study investigates Arabic word embedding models as alternatives to pre-trained models, resulting in the WordEmb.Collab-BiLSTM/CNN and WordEmb.Collab-BiLSTM/AVG models. Experimental results on our BARAQA (Big-ARAbic-Question-Answering) dataset and the SemEval Arabic Question-Answering corpus demonstrate that the AraELECTRA.Collab-BiLSTM/CNN model achieves high accuracies of 84.64% and 45.93%, respectively. Moreover, the WordEmb.Collab-BiLSTM/AVG model significantly enhances accuracy to 91.61% and 81.23% on the respective datasets, showcasing the effectiveness of our collaborative techniques. Our proposed architecture represents a substantial improvement over previous models, emphasizing the importance of advanced techniques and collaborative strategies in handling complex language structures and diverse text dependencies. Additionally, the study underscores the performance of Arabic transformer-based encoding and suggests further exploration of transformers and collaborative strategies to bolster AS2 performance.</p>","PeriodicalId":8109,"journal":{"name":"Arabian Journal for Science and Engineering","volume":null,"pages":null},"PeriodicalIF":2.9000,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"CollabAS2: Enhancing Arabic Answer Sentence Selection Using Transformer-Based Collaborative Models\",\"authors\":\"Asma Aouichat, Ahmed Guessoum\",\"doi\":\"10.1007/s13369-024-09345-3\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Accurately identifying pertinent text segments as answers to questions is crucial for optimizing question-answering systems, underscoring the pivotal role of precision in Answer Sentence Selection (AS2) modules. This study introduces an innovative AS2 module design leveraging the AraBERT transformer to encode inputs-one for the question and one for the candidate answer-with the goal of enhancing comprehension of both inputs. Each encoded input is subsequently processed in parallel by a collaborative layer employing two distinct deep learning models: a bidirectional long short-term memory (BiLSTM) and a convolutional neural network (CNN). This collaborative approach forms the AraBERT.Collab-BiLSTM/CNN model. Additionally, extensions to the study include AraBERT.Collab-BiLSTM/AVG, incorporating a BiLSTM and AVG collaboration layer, as well as the use of the AraELECTRA pre-trained model, yielding the AraELECTRA.Collab-BiLSTM/CNN and AraELECTRA.Collab-BiLSTM/AVG configurations. Furthermore, the study investigates Arabic word embedding models as alternatives to pre-trained models, resulting in the WordEmb.Collab-BiLSTM/CNN and WordEmb.Collab-BiLSTM/AVG models. Experimental results on our BARAQA (Big-ARAbic-Question-Answering) dataset and the SemEval Arabic Question-Answering corpus demonstrate that the AraELECTRA.Collab-BiLSTM/CNN model achieves high accuracies of 84.64% and 45.93%, respectively. Moreover, the WordEmb.Collab-BiLSTM/AVG model significantly enhances accuracy to 91.61% and 81.23% on the respective datasets, showcasing the effectiveness of our collaborative techniques. Our proposed architecture represents a substantial improvement over previous models, emphasizing the importance of advanced techniques and collaborative strategies in handling complex language structures and diverse text dependencies. Additionally, the study underscores the performance of Arabic transformer-based encoding and suggests further exploration of transformers and collaborative strategies to bolster AS2 performance.</p>\",\"PeriodicalId\":8109,\"journal\":{\"name\":\"Arabian Journal for Science and Engineering\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":2.9000,\"publicationDate\":\"2024-08-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Arabian Journal for Science and Engineering\",\"FirstCategoryId\":\"103\",\"ListUrlMain\":\"https://doi.org/10.1007/s13369-024-09345-3\",\"RegionNum\":4,\"RegionCategory\":\"综合性期刊\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"Multidisciplinary\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Arabian Journal for Science and Engineering","FirstCategoryId":"103","ListUrlMain":"https://doi.org/10.1007/s13369-024-09345-3","RegionNum":4,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"Multidisciplinary","Score":null,"Total":0}
引用次数: 0

摘要

准确识别作为问题答案的相关文本片段对于优化问题解答系统至关重要,这也凸显了答案句子选择(AS2)模块中精确性的关键作用。本研究介绍了一种创新的 AS2 模块设计,利用 AraBERT 转换器对输入进行编码,一个是问题输入,一个是候选答案输入,目的是提高对两个输入的理解能力。每个编码输入随后由一个协作层并行处理,协作层采用两种不同的深度学习模型:双向长短期记忆(BiLSTM)和卷积神经网络(CNN)。这种协作方法形成了 AraBERT.Collab-BiLSTM/CNN 模型。此外,该研究的扩展包括 AraBERT.Collab-BiLSTM/AVG,其中包含一个 BiLSTM 和 AVG 协作层,以及使用 AraELECTRA 预训练模型,产生 AraELECTRA.Collab-BiLSTM/CN 和 AraELECTRA.Collab-BiLSTM/AVG 配置。此外,该研究还调查了阿拉伯语单词嵌入模型,作为预训练模型的替代方案,从而产生了 WordEmb.Collab-BiLSTM/CN 和 WordEmb.Collab-BiLSTM/AVG 模型。在我们的 BARAQA(Big-ARAbic-Question-Answering)数据集和 SemEval 阿拉伯语问答语料库上的实验结果表明,AraELECTRA.Collab-BiLSTM/CNN 模型分别达到了 84.64% 和 45.93% 的高准确率。此外,WordEmb.Collab-BiLSTM/AVG 模型也显著提高了各自数据集的准确率,分别达到 91.61% 和 81.23%,展示了我们协作技术的有效性。与以前的模型相比,我们提出的架构有了很大的改进,强调了先进技术和协作策略在处理复杂语言结构和多样化文本依赖关系方面的重要性。此外,这项研究还强调了基于转换器的阿拉伯语编码的性能,并建议进一步探索转换器和协作策略,以提高 AS2 的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

CollabAS2: Enhancing Arabic Answer Sentence Selection Using Transformer-Based Collaborative Models

CollabAS2: Enhancing Arabic Answer Sentence Selection Using Transformer-Based Collaborative Models

Accurately identifying pertinent text segments as answers to questions is crucial for optimizing question-answering systems, underscoring the pivotal role of precision in Answer Sentence Selection (AS2) modules. This study introduces an innovative AS2 module design leveraging the AraBERT transformer to encode inputs-one for the question and one for the candidate answer-with the goal of enhancing comprehension of both inputs. Each encoded input is subsequently processed in parallel by a collaborative layer employing two distinct deep learning models: a bidirectional long short-term memory (BiLSTM) and a convolutional neural network (CNN). This collaborative approach forms the AraBERT.Collab-BiLSTM/CNN model. Additionally, extensions to the study include AraBERT.Collab-BiLSTM/AVG, incorporating a BiLSTM and AVG collaboration layer, as well as the use of the AraELECTRA pre-trained model, yielding the AraELECTRA.Collab-BiLSTM/CNN and AraELECTRA.Collab-BiLSTM/AVG configurations. Furthermore, the study investigates Arabic word embedding models as alternatives to pre-trained models, resulting in the WordEmb.Collab-BiLSTM/CNN and WordEmb.Collab-BiLSTM/AVG models. Experimental results on our BARAQA (Big-ARAbic-Question-Answering) dataset and the SemEval Arabic Question-Answering corpus demonstrate that the AraELECTRA.Collab-BiLSTM/CNN model achieves high accuracies of 84.64% and 45.93%, respectively. Moreover, the WordEmb.Collab-BiLSTM/AVG model significantly enhances accuracy to 91.61% and 81.23% on the respective datasets, showcasing the effectiveness of our collaborative techniques. Our proposed architecture represents a substantial improvement over previous models, emphasizing the importance of advanced techniques and collaborative strategies in handling complex language structures and diverse text dependencies. Additionally, the study underscores the performance of Arabic transformer-based encoding and suggests further exploration of transformers and collaborative strategies to bolster AS2 performance.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Arabian Journal for Science and Engineering
Arabian Journal for Science and Engineering 综合性期刊-综合性期刊
CiteScore
5.20
自引率
3.40%
发文量
0
审稿时长
4.3 months
期刊介绍: King Fahd University of Petroleum & Minerals (KFUPM) partnered with Springer to publish the Arabian Journal for Science and Engineering (AJSE). AJSE, which has been published by KFUPM since 1975, is a recognized national, regional and international journal that provides a great opportunity for the dissemination of research advances from the Kingdom of Saudi Arabia, MENA and the world.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信