A deep neural architecture for sentence semantic matching

Xu Zhang, Wenpeng Lu, Fangfang Li, Ruoyu Zhang, Jinyong Cheng
{"title":"A deep neural architecture for sentence semantic matching","authors":"Xu Zhang, Wenpeng Lu, Fangfang Li, Ruoyu Zhang, Jinyong Cheng","doi":"10.1504/ijcse.2020.10028622","DOIUrl":null,"url":null,"abstract":"Sentence semantic matching (SSM) is a fundamental research task in natural language processing. Most existing SSM methods take the advantage of sentence representation learning to generate a single or multi-granularity semantic representation for sentence matching. However, sentence interactions and loss function which are the two key factors for SSM still have not been fully considered. Accordingly, we propose a deep neural network architecture for SSM task with a sentence interactive matching layer and an optimised loss function. Given two input sentences, our model first encodes them to embeddings with an ordinary long short-term memory (LSTM) encoder. Then, the encoded embeddings are handled by an attention layer to find the key and important words in the sentences. Next, sentence interactions are captured with a matching layer to output a matching vector. Finally, based on the matching vector, a fully connected multi-layer perceptron outputs the similarity score. The model also distinguishes the equivocation training instances with an improved optimised loss function. We also systematically evaluate our model on a public Chinese semantic matching corpus, BQ corpus. The results demonstrate that our model outperforms the state-of-the-art methods, i.e., BiMPM, DIIN.","PeriodicalId":340410,"journal":{"name":"Int. J. Comput. Sci. Eng.","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Int. J. Comput. Sci. Eng.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1504/ijcse.2020.10028622","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

Abstract

Sentence semantic matching (SSM) is a fundamental research task in natural language processing. Most existing SSM methods take the advantage of sentence representation learning to generate a single or multi-granularity semantic representation for sentence matching. However, sentence interactions and loss function which are the two key factors for SSM still have not been fully considered. Accordingly, we propose a deep neural network architecture for SSM task with a sentence interactive matching layer and an optimised loss function. Given two input sentences, our model first encodes them to embeddings with an ordinary long short-term memory (LSTM) encoder. Then, the encoded embeddings are handled by an attention layer to find the key and important words in the sentences. Next, sentence interactions are captured with a matching layer to output a matching vector. Finally, based on the matching vector, a fully connected multi-layer perceptron outputs the similarity score. The model also distinguishes the equivocation training instances with an improved optimised loss function. We also systematically evaluate our model on a public Chinese semantic matching corpus, BQ corpus. The results demonstrate that our model outperforms the state-of-the-art methods, i.e., BiMPM, DIIN.
一种用于句子语义匹配的深度神经结构
句子语义匹配是自然语言处理领域的一项基础性研究课题。大多数现有的SSM方法都利用句子表示学习的优势,为句子匹配生成单粒度或多粒度的语义表示。然而,作为SSM的两个关键因素,句子相互作用和损失函数仍然没有得到充分的考虑。因此,我们提出了一种基于句子交互匹配层和优化损失函数的深度神经网络结构。给定两个输入句子,我们的模型首先用一个普通的长短期记忆(LSTM)编码器将它们编码为嵌入。然后,通过注意层处理编码后的嵌入,找到句子中的关键字和重要词。接下来,用匹配层捕获句子交互以输出匹配向量。最后,基于匹配向量,一个完全连接的多层感知器输出相似度得分。该模型还使用改进的优化损失函数来区分歧义训练实例。我们还在一个公共汉语语义匹配语料库(BQ语料库)上系统地评估了我们的模型。结果表明,我们的模型优于最先进的方法,即BiMPM, DIIN。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信