基于两个独立通道的文本相似度:暹罗卷积神经网络和暹罗循环神经网络

IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Zhengfang He
{"title":"基于两个独立通道的文本相似度:暹罗卷积神经网络和暹罗循环神经网络","authors":"Zhengfang He","doi":"10.1016/j.neucom.2025.130355","DOIUrl":null,"url":null,"abstract":"<div><div>In the present-day context, a large amount of information exists in text. It is hard to extract meaningful and potential information from the text. From the current research, text similarity provides a method applied in many practical scenarios. Traditional text similarity algorithms are easy to implement, but more than these algorithms are needed to extract text features. At present, most text similarity algorithms are based on deep learning. However, these algorithms often struggle with adequately extracting both local and context text features, and they typically do not differentiate between the effectiveness of these two types of feature extraction. To address these shortcomings, this paper proposes Two Independent Channels: Siamese Convolutional Neural Networks and Siamese Recurrent Neural Networks (TIC-SCNN-SRNN). This approach is designed for binary classification, where ‘1’ indicates similarity and ‘0’ indicates dissimilarity. In detail, this paper proposes the Siamese Convolutional Neural Networks (SCNN) model to address the issue of insufficient extraction of local text features. Additionally, it introduces the Siamese Recurrent Neural Networks (SRNN) model to tackle the problem of insufficient extraction of context text features. Due to the issue of not distinguishing the effects of local and context text features extractions, this paper conducts independent weight learning on these two models to research which is more effective for text similarity tasks. In order to verify the effectiveness of the model, this paper experiments on SciTail, TwitterPPD, and QQP datasets. The experimental results show that SCNN and SRNN influence the text similarity tasks, but SRNN is more effective than SCNN. Furthermore, to verify the advantages of the TIC-SCNN-SRNN model, this paper tests the performance of the state-of-the-art CNN&amp;RNN models. The test results show that the TIC-SCNN-SRNN model performs the best, indicating that the model proposed in this paper is more effective for the text similarity tasks.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"643 ","pages":"Article 130355"},"PeriodicalIF":5.5000,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Text similarity based on two independent channels: Siamese Convolutional Neural Networks and Siamese Recurrent Neural Networks\",\"authors\":\"Zhengfang He\",\"doi\":\"10.1016/j.neucom.2025.130355\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>In the present-day context, a large amount of information exists in text. It is hard to extract meaningful and potential information from the text. From the current research, text similarity provides a method applied in many practical scenarios. Traditional text similarity algorithms are easy to implement, but more than these algorithms are needed to extract text features. At present, most text similarity algorithms are based on deep learning. However, these algorithms often struggle with adequately extracting both local and context text features, and they typically do not differentiate between the effectiveness of these two types of feature extraction. To address these shortcomings, this paper proposes Two Independent Channels: Siamese Convolutional Neural Networks and Siamese Recurrent Neural Networks (TIC-SCNN-SRNN). This approach is designed for binary classification, where ‘1’ indicates similarity and ‘0’ indicates dissimilarity. In detail, this paper proposes the Siamese Convolutional Neural Networks (SCNN) model to address the issue of insufficient extraction of local text features. Additionally, it introduces the Siamese Recurrent Neural Networks (SRNN) model to tackle the problem of insufficient extraction of context text features. Due to the issue of not distinguishing the effects of local and context text features extractions, this paper conducts independent weight learning on these two models to research which is more effective for text similarity tasks. In order to verify the effectiveness of the model, this paper experiments on SciTail, TwitterPPD, and QQP datasets. The experimental results show that SCNN and SRNN influence the text similarity tasks, but SRNN is more effective than SCNN. Furthermore, to verify the advantages of the TIC-SCNN-SRNN model, this paper tests the performance of the state-of-the-art CNN&amp;RNN models. The test results show that the TIC-SCNN-SRNN model performs the best, indicating that the model proposed in this paper is more effective for the text similarity tasks.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"643 \",\"pages\":\"Article 130355\"},\"PeriodicalIF\":5.5000,\"publicationDate\":\"2025-05-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225010276\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225010276","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

在当今的语境中,大量的信息存在于语篇中。很难从文本中提取有意义和潜在的信息。从目前的研究来看,文本相似度提供了一种应用于许多实际场景的方法。传统的文本相似度算法很容易实现,但提取文本特征需要更多的算法。目前,大多数文本相似度算法都是基于深度学习的。然而,这些算法往往难以充分提取局部和上下文文本特征,而且它们通常不区分这两种类型的特征提取的有效性。为了解决这些缺点,本文提出了两个独立的通道:暹罗卷积神经网络和暹罗循环神经网络(TIC-SCNN-SRNN)。这种方法是为二元分类设计的,其中“1”表示相似,“0”表示不相似。本文提出了Siamese卷积神经网络(SCNN)模型来解决局部文本特征提取不足的问题。此外,还引入了连体递归神经网络(Siamese Recurrent Neural Networks, SRNN)模型来解决上下文文本特征提取不足的问题。由于没有区分局部和上下文文本特征提取的效果,本文对这两种模型进行了独立的权值学习,以研究哪种模型对文本相似任务更有效。为了验证模型的有效性,本文在SciTail、TwitterPPD和QQP数据集上进行了实验。实验结果表明,SCNN和SRNN对文本相似度任务都有影响,但SRNN比SCNN更有效。此外,为了验证TIC-SCNN-SRNN模型的优势,本文测试了最先进的CNN&;RNN模型的性能。测试结果表明,TIC-SCNN-SRNN模型表现最好,表明本文提出的模型对于文本相似度任务更为有效。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Text similarity based on two independent channels: Siamese Convolutional Neural Networks and Siamese Recurrent Neural Networks
In the present-day context, a large amount of information exists in text. It is hard to extract meaningful and potential information from the text. From the current research, text similarity provides a method applied in many practical scenarios. Traditional text similarity algorithms are easy to implement, but more than these algorithms are needed to extract text features. At present, most text similarity algorithms are based on deep learning. However, these algorithms often struggle with adequately extracting both local and context text features, and they typically do not differentiate between the effectiveness of these two types of feature extraction. To address these shortcomings, this paper proposes Two Independent Channels: Siamese Convolutional Neural Networks and Siamese Recurrent Neural Networks (TIC-SCNN-SRNN). This approach is designed for binary classification, where ‘1’ indicates similarity and ‘0’ indicates dissimilarity. In detail, this paper proposes the Siamese Convolutional Neural Networks (SCNN) model to address the issue of insufficient extraction of local text features. Additionally, it introduces the Siamese Recurrent Neural Networks (SRNN) model to tackle the problem of insufficient extraction of context text features. Due to the issue of not distinguishing the effects of local and context text features extractions, this paper conducts independent weight learning on these two models to research which is more effective for text similarity tasks. In order to verify the effectiveness of the model, this paper experiments on SciTail, TwitterPPD, and QQP datasets. The experimental results show that SCNN and SRNN influence the text similarity tasks, but SRNN is more effective than SCNN. Furthermore, to verify the advantages of the TIC-SCNN-SRNN model, this paper tests the performance of the state-of-the-art CNN&RNN models. The test results show that the TIC-SCNN-SRNN model performs the best, indicating that the model proposed in this paper is more effective for the text similarity tasks.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信