{"title":"BRFP:一种支持融合语法和图嵌入表示算法的高效通用的句子嵌入学习模型方法","authors":"Zhifeng Li, Wen-Wang Wu, Chunlei Shen","doi":"10.1155/2022/7471408","DOIUrl":null,"url":null,"abstract":"Due to the rapidly growing volume of data on the Internet, the methods of efficiently and accurately processing massive text information have been the focus of research. In natural language processing theory, sentence embedding representation is an important method. This paper proposes a new sentence embedding learning model called BRFP (Factorization Process with Bidirectional Restraints) that fuses syntactic information, uses matrix decomposition to learn syntactic information, and fuses and calculates with word vectors to obtain the embedded representation of sentences. In the experimental chapter, text similarity experiments are conducted to verify the rationality and effectiveness of the model and analyzed experimental results on Chinese and English texts with the current mainstream learning methods, and potential improvement directions are summarized. The experimental results on Chinese and English datasets, including STS, AFQMC, and LCQMC, show that the model proposed in this paper outperforms the CNN method in terms of accuracy and F1 value by 7.6% and 4.8. The comparison experiment with the word vector weighted model shows that when the sentence length is longer, or the corresponding syntactic structure is complex, the model’s advantages in this paper are more prominent than TF-IDF and SIF methods. Compared with the TF-IDF method, the effect improved by 14.4%. Compared with the SIF method, it has a maximum advantage of 7.9%, and the overall improvement in each comparative experimental task is between 4 and 6 percentage points. In the neural network model comparison experiment, the model in this paper compared the CNN, RNN, LSTM, ST, QT, and InferSent models, and the effect significantly improved on the 14’OnWN, 14’Tweet-news, and 15’Ans.-forum datasets. For example, in the 14’OnWN dataset, the BRFP method has a 10.9% improvement over the ST method. The 14’Tweet-news dataset has a 22.9% advantage over the LSTM method, and the 15’Ans.-forum dataset has a 24.07% improvement over the RNN method. The article also demonstrates the generality of the model, proving that the model proposed in this paper is also a universal learning framework.","PeriodicalId":14776,"journal":{"name":"J. Sensors","volume":"03 1","pages":"1-14"},"PeriodicalIF":0.0000,"publicationDate":"2022-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"BRFP: An Efficient and Universal Sentence Embedding Learning Model Method Supporting Fused Syntax Combined with Graph Embedding Representation Algorithm\",\"authors\":\"Zhifeng Li, Wen-Wang Wu, Chunlei Shen\",\"doi\":\"10.1155/2022/7471408\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Due to the rapidly growing volume of data on the Internet, the methods of efficiently and accurately processing massive text information have been the focus of research. In natural language processing theory, sentence embedding representation is an important method. This paper proposes a new sentence embedding learning model called BRFP (Factorization Process with Bidirectional Restraints) that fuses syntactic information, uses matrix decomposition to learn syntactic information, and fuses and calculates with word vectors to obtain the embedded representation of sentences. In the experimental chapter, text similarity experiments are conducted to verify the rationality and effectiveness of the model and analyzed experimental results on Chinese and English texts with the current mainstream learning methods, and potential improvement directions are summarized. The experimental results on Chinese and English datasets, including STS, AFQMC, and LCQMC, show that the model proposed in this paper outperforms the CNN method in terms of accuracy and F1 value by 7.6% and 4.8. The comparison experiment with the word vector weighted model shows that when the sentence length is longer, or the corresponding syntactic structure is complex, the model’s advantages in this paper are more prominent than TF-IDF and SIF methods. Compared with the TF-IDF method, the effect improved by 14.4%. Compared with the SIF method, it has a maximum advantage of 7.9%, and the overall improvement in each comparative experimental task is between 4 and 6 percentage points. In the neural network model comparison experiment, the model in this paper compared the CNN, RNN, LSTM, ST, QT, and InferSent models, and the effect significantly improved on the 14’OnWN, 14’Tweet-news, and 15’Ans.-forum datasets. For example, in the 14’OnWN dataset, the BRFP method has a 10.9% improvement over the ST method. The 14’Tweet-news dataset has a 22.9% advantage over the LSTM method, and the 15’Ans.-forum dataset has a 24.07% improvement over the RNN method. The article also demonstrates the generality of the model, proving that the model proposed in this paper is also a universal learning framework.\",\"PeriodicalId\":14776,\"journal\":{\"name\":\"J. Sensors\",\"volume\":\"03 1\",\"pages\":\"1-14\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-08-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"J. Sensors\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1155/2022/7471408\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"J. Sensors","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1155/2022/7471408","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
随着互联网数据量的快速增长,如何高效、准确地处理海量文本信息一直是研究的热点。在自然语言处理理论中,句子嵌入表示是一种重要的方法。本文提出了一种新的句子嵌入学习模型BRFP (Factorization Process with Bidirectional constraints),该模型融合句法信息,利用矩阵分解学习句法信息,再结合词向量进行融合计算,得到句子的嵌入表示。实验章节通过文本相似度实验验证了模型的合理性和有效性,并对当前主流学习方法下的中英文文本实验结果进行了分析,总结了可能的改进方向。在STS、AFQMC和题中英文数据集上的实验结果表明,本文提出的模型在准确率和F1值上分别优于CNN方法7.6%和4.8。与词向量加权模型的对比实验表明,当句子长度较长或对应的句法结构较复杂时,本文模型的优势比TF-IDF和SIF方法更为突出。与TF-IDF法相比,效果提高14.4%。与SIF方法相比,其最大优势为7.9%,各对比实验任务的整体提升幅度在4 - 6个百分点之间。在神经网络模型对比实验中,本文模型对比了CNN、RNN、LSTM、ST、QT和InferSent模型,在14'OnWN、14'Tweet-news和15'Ans上的效果显著提高。论坛数据集。例如,在14'OnWN数据集中,BRFP方法比ST方法有10.9%的改进。14'Tweet-news数据集比LSTM方法和15'Ans方法具有22.9%的优势。-forum数据集比RNN方法提高了24.07%。文章还论证了模型的通用性,证明本文提出的模型也是一个通用的学习框架。
BRFP: An Efficient and Universal Sentence Embedding Learning Model Method Supporting Fused Syntax Combined with Graph Embedding Representation Algorithm
Due to the rapidly growing volume of data on the Internet, the methods of efficiently and accurately processing massive text information have been the focus of research. In natural language processing theory, sentence embedding representation is an important method. This paper proposes a new sentence embedding learning model called BRFP (Factorization Process with Bidirectional Restraints) that fuses syntactic information, uses matrix decomposition to learn syntactic information, and fuses and calculates with word vectors to obtain the embedded representation of sentences. In the experimental chapter, text similarity experiments are conducted to verify the rationality and effectiveness of the model and analyzed experimental results on Chinese and English texts with the current mainstream learning methods, and potential improvement directions are summarized. The experimental results on Chinese and English datasets, including STS, AFQMC, and LCQMC, show that the model proposed in this paper outperforms the CNN method in terms of accuracy and F1 value by 7.6% and 4.8. The comparison experiment with the word vector weighted model shows that when the sentence length is longer, or the corresponding syntactic structure is complex, the model’s advantages in this paper are more prominent than TF-IDF and SIF methods. Compared with the TF-IDF method, the effect improved by 14.4%. Compared with the SIF method, it has a maximum advantage of 7.9%, and the overall improvement in each comparative experimental task is between 4 and 6 percentage points. In the neural network model comparison experiment, the model in this paper compared the CNN, RNN, LSTM, ST, QT, and InferSent models, and the effect significantly improved on the 14’OnWN, 14’Tweet-news, and 15’Ans.-forum datasets. For example, in the 14’OnWN dataset, the BRFP method has a 10.9% improvement over the ST method. The 14’Tweet-news dataset has a 22.9% advantage over the LSTM method, and the 15’Ans.-forum dataset has a 24.07% improvement over the RNN method. The article also demonstrates the generality of the model, proving that the model proposed in this paper is also a universal learning framework.