A Text Summarization Model with Enhanced Local Relevance

Ouyang Ning, Junyan Wang, Peng Jiang, Xiao-Sheng Cai
{"title":"A Text Summarization Model with Enhanced Local Relevance","authors":"Ouyang Ning, Junyan Wang, Peng Jiang, Xiao-Sheng Cai","doi":"10.1109/IAEAC54830.2022.9929557","DOIUrl":null,"url":null,"abstract":"When generating short text summaries., it is challenging to accurately learn the global semantic information of the original text and extract the correlation features between local semantic information., and at the same time lead to too much redundant information., making the generated summaries ineffective. In addition., the existing normalization algorithm will increase the computational complexity of the text summarization model., which affects the performance of the model. Aiming at the above problems., a text summarization generation model GMELC(Generation Model for Enhancing Local Correlation) is proposed to enhance local correlation in generated summaries. First., the residual concept used in other media feature extraction networks is introduced into the text summarization model. We add the word semantic feature as a residual block to the n-gram feature., which improves the dependencies of words in phrases and strengthens the correlation between phrases and words in sentences. Secondly., we propose a scaled I2 normalization method to normalize the data for reducing the amount of training parameters and removing the unnecessary computation caused by variance., so that the computational complexity of the model is reduced., thereby improving the computational efficiency and performance of the model. In order to verify the role of the model in enhancing the correlation between Chinese characters and words., experiments were conducted on the Chinese dataset LCSTS., the result shows that the summaries generated by GMELC have higher recall and better readability than other state-of-the-art models.","PeriodicalId":349113,"journal":{"name":"2022 IEEE 6th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC )","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 6th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC )","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IAEAC54830.2022.9929557","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

When generating short text summaries., it is challenging to accurately learn the global semantic information of the original text and extract the correlation features between local semantic information., and at the same time lead to too much redundant information., making the generated summaries ineffective. In addition., the existing normalization algorithm will increase the computational complexity of the text summarization model., which affects the performance of the model. Aiming at the above problems., a text summarization generation model GMELC(Generation Model for Enhancing Local Correlation) is proposed to enhance local correlation in generated summaries. First., the residual concept used in other media feature extraction networks is introduced into the text summarization model. We add the word semantic feature as a residual block to the n-gram feature., which improves the dependencies of words in phrases and strengthens the correlation between phrases and words in sentences. Secondly., we propose a scaled I2 normalization method to normalize the data for reducing the amount of training parameters and removing the unnecessary computation caused by variance., so that the computational complexity of the model is reduced., thereby improving the computational efficiency and performance of the model. In order to verify the role of the model in enhancing the correlation between Chinese characters and words., experiments were conducted on the Chinese dataset LCSTS., the result shows that the summaries generated by GMELC have higher recall and better readability than other state-of-the-art models.
一种增强局部相关性的文本摘要模型
在生成简短的文本摘要时。,如何准确地学习原始文本的全局语义信息和提取局部语义信息之间的关联特征是一个挑战。,同时也会产生过多的冗余信息。,使得生成的摘要无效。此外。现有的归一化算法会增加文本摘要模型的计算复杂度。,这会影响模型的性能。针对以上问题。提出了一种文本摘要生成模型GMELC(generation model for enhancements Local Correlation),用于增强生成摘要中的局部相关性。第一。,将其他媒体特征提取网络中使用的残差概念引入文本摘要模型。我们将单词语义特征作为残差块添加到n-gram特征中。,提高了短语中单词的依赖性,加强了短语与句子中单词的相关性。其次。,我们提出了一种缩放I2归一化方法对数据进行归一化,以减少训练参数的数量,并消除方差带来的不必要的计算。,从而降低了模型的计算复杂度。,从而提高模型的计算效率和性能。为了验证该模型在增强汉字与词的相关性方面的作用。,在中文数据集LCSTS上进行了实验。结果表明,GMELC生成的摘要具有更高的查全率和更好的可读性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信