HASumRuNNer: An Extractive Text Summarization Optimization Model Based on a Gradient-Based Algorithm

IF 0.9 Q4 COMPUTER SCIENCE, INFORMATION SYSTEMS
Muljono, M. Nababan, R. A. Nugroho, Kevin Djajadinata
{"title":"HASumRuNNer: An Extractive Text Summarization Optimization Model Based on a Gradient-Based Algorithm","authors":"Muljono, M. Nababan, R. A. Nugroho, Kevin Djajadinata","doi":"10.12720/jait.14.4.656-667","DOIUrl":null,"url":null,"abstract":"—This article is based on text summarization research model, also referred to as “text summarization”, which is the act of summarizing materials in a way that directly communicates the intent or message of a document. Hierarchical Attention SumRuNNer (HASumRuNNer), an extractive text summary model based on the Indonesian language is the text summary model suggested in this study. This is a novelty for the extractive text summary model based on the Indonesian language, as there is currently very few related research, both in terms of the approach and dataset. Three primary methods—BiGRU, CharCNN, and hierarchical attention mechanisms—were used to create the model for this study. The optimization in this suggested model is likewise carried out using a variety of gradient-based methods, and the ROUGE-N approach is used to assess the outcomes of text synthesis. The test results demonstrate that Adam’s gradient-based approach is the most effective for extracting text summarization using the HASumRuNNer model. As can be seen, the values of RED-1 (70.7), RED-2 (64.33), and RED-L (68.14) are greater than those of other methods employed as references. The approach used in the suggested HASumRuNNer Model, which combines BiGRU with CharCNN, can result in more accurate word and sentence representations at word and sentence levels. Additionally, the word and sentence-level hierarchical attention mechanisms aid in preventing the loss of information on each word in documents that are typically brought on by the length of the input model word or sentence.","PeriodicalId":36452,"journal":{"name":"Journal of Advances in Information Technology","volume":"1 1","pages":""},"PeriodicalIF":0.9000,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Advances in Information Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.12720/jait.14.4.656-667","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

—This article is based on text summarization research model, also referred to as “text summarization”, which is the act of summarizing materials in a way that directly communicates the intent or message of a document. Hierarchical Attention SumRuNNer (HASumRuNNer), an extractive text summary model based on the Indonesian language is the text summary model suggested in this study. This is a novelty for the extractive text summary model based on the Indonesian language, as there is currently very few related research, both in terms of the approach and dataset. Three primary methods—BiGRU, CharCNN, and hierarchical attention mechanisms—were used to create the model for this study. The optimization in this suggested model is likewise carried out using a variety of gradient-based methods, and the ROUGE-N approach is used to assess the outcomes of text synthesis. The test results demonstrate that Adam’s gradient-based approach is the most effective for extracting text summarization using the HASumRuNNer model. As can be seen, the values of RED-1 (70.7), RED-2 (64.33), and RED-L (68.14) are greater than those of other methods employed as references. The approach used in the suggested HASumRuNNer Model, which combines BiGRU with CharCNN, can result in more accurate word and sentence representations at word and sentence levels. Additionally, the word and sentence-level hierarchical attention mechanisms aid in preventing the loss of information on each word in documents that are typically brought on by the length of the input model word or sentence.
基于梯度算法的提取文本摘要优化模型HASumRuNNer
-本文基于文本摘要研究模型,也称为“文本摘要”,是一种以直接传达文档意图或信息的方式对材料进行总结的行为。本文提出了基于印尼语的文本摘要抽取模型——层次注意SumRuNNer (HASumRuNNer)。这是基于印尼语的提取文本摘要模型的新颖之处,因为目前相关的研究很少,无论是在方法还是数据集方面。本研究采用了bigru、CharCNN和分层注意机制三种主要方法来创建模型。该模型的优化同样使用各种基于梯度的方法进行,并使用ROUGE-N方法来评估文本合成的结果。测试结果表明,Adam的基于梯度的方法对于使用HASumRuNNer模型提取文本摘要是最有效的。可以看出,RED-1(70.7)、RED-2(64.33)和RED-L(68.14)的值均大于其他参考方法。在建议的HASumRuNNer模型中使用的方法将BiGRU与CharCNN结合起来,可以在单词和句子级别上产生更准确的单词和句子表示。此外,单词和句子级别的分层注意机制有助于防止文档中每个单词的信息丢失,这种丢失通常是由输入模型单词或句子的长度引起的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Journal of Advances in Information Technology
Journal of Advances in Information Technology Computer Science-Information Systems
CiteScore
4.20
自引率
20.00%
发文量
46
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信