Ouyang Ning, Junyan Wang, Peng Jiang, Xiao-Sheng Cai
{"title":"一种增强局部相关性的文本摘要模型","authors":"Ouyang Ning, Junyan Wang, Peng Jiang, Xiao-Sheng Cai","doi":"10.1109/IAEAC54830.2022.9929557","DOIUrl":null,"url":null,"abstract":"When generating short text summaries., it is challenging to accurately learn the global semantic information of the original text and extract the correlation features between local semantic information., and at the same time lead to too much redundant information., making the generated summaries ineffective. In addition., the existing normalization algorithm will increase the computational complexity of the text summarization model., which affects the performance of the model. Aiming at the above problems., a text summarization generation model GMELC(Generation Model for Enhancing Local Correlation) is proposed to enhance local correlation in generated summaries. First., the residual concept used in other media feature extraction networks is introduced into the text summarization model. We add the word semantic feature as a residual block to the n-gram feature., which improves the dependencies of words in phrases and strengthens the correlation between phrases and words in sentences. Secondly., we propose a scaled I2 normalization method to normalize the data for reducing the amount of training parameters and removing the unnecessary computation caused by variance., so that the computational complexity of the model is reduced., thereby improving the computational efficiency and performance of the model. In order to verify the role of the model in enhancing the correlation between Chinese characters and words., experiments were conducted on the Chinese dataset LCSTS., the result shows that the summaries generated by GMELC have higher recall and better readability than other state-of-the-art models.","PeriodicalId":349113,"journal":{"name":"2022 IEEE 6th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC )","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Text Summarization Model with Enhanced Local Relevance\",\"authors\":\"Ouyang Ning, Junyan Wang, Peng Jiang, Xiao-Sheng Cai\",\"doi\":\"10.1109/IAEAC54830.2022.9929557\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"When generating short text summaries., it is challenging to accurately learn the global semantic information of the original text and extract the correlation features between local semantic information., and at the same time lead to too much redundant information., making the generated summaries ineffective. In addition., the existing normalization algorithm will increase the computational complexity of the text summarization model., which affects the performance of the model. Aiming at the above problems., a text summarization generation model GMELC(Generation Model for Enhancing Local Correlation) is proposed to enhance local correlation in generated summaries. First., the residual concept used in other media feature extraction networks is introduced into the text summarization model. We add the word semantic feature as a residual block to the n-gram feature., which improves the dependencies of words in phrases and strengthens the correlation between phrases and words in sentences. Secondly., we propose a scaled I2 normalization method to normalize the data for reducing the amount of training parameters and removing the unnecessary computation caused by variance., so that the computational complexity of the model is reduced., thereby improving the computational efficiency and performance of the model. In order to verify the role of the model in enhancing the correlation between Chinese characters and words., experiments were conducted on the Chinese dataset LCSTS., the result shows that the summaries generated by GMELC have higher recall and better readability than other state-of-the-art models.\",\"PeriodicalId\":349113,\"journal\":{\"name\":\"2022 IEEE 6th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC )\",\"volume\":\"11 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-10-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 6th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC )\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IAEAC54830.2022.9929557\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 6th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC )","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IAEAC54830.2022.9929557","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
在生成简短的文本摘要时。,如何准确地学习原始文本的全局语义信息和提取局部语义信息之间的关联特征是一个挑战。,同时也会产生过多的冗余信息。,使得生成的摘要无效。此外。现有的归一化算法会增加文本摘要模型的计算复杂度。,这会影响模型的性能。针对以上问题。提出了一种文本摘要生成模型GMELC(generation model for enhancements Local Correlation),用于增强生成摘要中的局部相关性。第一。,将其他媒体特征提取网络中使用的残差概念引入文本摘要模型。我们将单词语义特征作为残差块添加到n-gram特征中。,提高了短语中单词的依赖性,加强了短语与句子中单词的相关性。其次。,我们提出了一种缩放I2归一化方法对数据进行归一化,以减少训练参数的数量,并消除方差带来的不必要的计算。,从而降低了模型的计算复杂度。,从而提高模型的计算效率和性能。为了验证该模型在增强汉字与词的相关性方面的作用。,在中文数据集LCSTS上进行了实验。结果表明,GMELC生成的摘要具有更高的查全率和更好的可读性。
A Text Summarization Model with Enhanced Local Relevance
When generating short text summaries., it is challenging to accurately learn the global semantic information of the original text and extract the correlation features between local semantic information., and at the same time lead to too much redundant information., making the generated summaries ineffective. In addition., the existing normalization algorithm will increase the computational complexity of the text summarization model., which affects the performance of the model. Aiming at the above problems., a text summarization generation model GMELC(Generation Model for Enhancing Local Correlation) is proposed to enhance local correlation in generated summaries. First., the residual concept used in other media feature extraction networks is introduced into the text summarization model. We add the word semantic feature as a residual block to the n-gram feature., which improves the dependencies of words in phrases and strengthens the correlation between phrases and words in sentences. Secondly., we propose a scaled I2 normalization method to normalize the data for reducing the amount of training parameters and removing the unnecessary computation caused by variance., so that the computational complexity of the model is reduced., thereby improving the computational efficiency and performance of the model. In order to verify the role of the model in enhancing the correlation between Chinese characters and words., experiments were conducted on the Chinese dataset LCSTS., the result shows that the summaries generated by GMELC have higher recall and better readability than other state-of-the-art models.