Attention based Recurrent Neural Network for Nepali Text Summarization

Bipin Timalsina, N. Paudel, T. B. Shahi
{"title":"Attention based Recurrent Neural Network for Nepali Text Summarization","authors":"Bipin Timalsina, N. Paudel, T. B. Shahi","doi":"10.3126/jist.v27i1.46709","DOIUrl":null,"url":null,"abstract":"Automatic text summarization has been a challenging topic in natural language processing (NLP) as it demands preserving important information while summarizing the large text into a summary. Extractive and abstractive text summarization are widely investigated approaches for text summarization. In extractive summarization, the important sentence from the large text is extracted and combined to create a summary whereas abstractive summarization creates a summary that is more focused on meaning, rather than content. Therefore, abstractive summarization gained more attention from researchers in the recent past. However, text summarization is still an untouched topic in the Nepali language. To this end, we proposed an abstractive text summarization for Nepali text. Here, we, first, create a Nepali text dataset by scraping Nepali news from the online news portals. Second, we design a deep learning-based text summarization model based on an encoder-decoder recurrent neural network with attention. More precisely, Long Short-Term Memory (LSTM) cells are used in the encoder and decoder layer. Third, we build nine different models by selecting various hyper-parameters such as the number of hidden layers and the number of nodes. Finally, we report the Recall-Oriented Understudy for Gisting Evaluation (ROUGE) score for each model to evaluate their performance. Among nine different models created by adjusting different numbers of layers and hidden states, the model with a single-layer encoder and 256 hidden states outperformed all other models with F-Score values of 15.74, 3.29, and 15.21 for ROUGE-1 ROUGE-2 and ROUGE-L, respectively.","PeriodicalId":16072,"journal":{"name":"Journal of Hunan Institute of Science and Technology","volume":"6 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2022-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Hunan Institute of Science and Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3126/jist.v27i1.46709","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Automatic text summarization has been a challenging topic in natural language processing (NLP) as it demands preserving important information while summarizing the large text into a summary. Extractive and abstractive text summarization are widely investigated approaches for text summarization. In extractive summarization, the important sentence from the large text is extracted and combined to create a summary whereas abstractive summarization creates a summary that is more focused on meaning, rather than content. Therefore, abstractive summarization gained more attention from researchers in the recent past. However, text summarization is still an untouched topic in the Nepali language. To this end, we proposed an abstractive text summarization for Nepali text. Here, we, first, create a Nepali text dataset by scraping Nepali news from the online news portals. Second, we design a deep learning-based text summarization model based on an encoder-decoder recurrent neural network with attention. More precisely, Long Short-Term Memory (LSTM) cells are used in the encoder and decoder layer. Third, we build nine different models by selecting various hyper-parameters such as the number of hidden layers and the number of nodes. Finally, we report the Recall-Oriented Understudy for Gisting Evaluation (ROUGE) score for each model to evaluate their performance. Among nine different models created by adjusting different numbers of layers and hidden states, the model with a single-layer encoder and 256 hidden states outperformed all other models with F-Score values of 15.74, 3.29, and 15.21 for ROUGE-1 ROUGE-2 and ROUGE-L, respectively.
基于注意的递归神经网络尼泊尔语文本摘要
自动文本摘要是自然语言处理(NLP)中的一个具有挑战性的课题,因为它要求在将大量文本汇总为摘要的同时保留重要信息。摘要抽取和抽象是目前被广泛研究的文本摘要方法。在抽象化的总结中,从大量的文本中提取出重要的句子并组合成一个总结,而抽象化的总结更注重意义,而不是内容。因此,近年来抽象总结越来越受到研究者的关注。然而,文本摘要在尼泊尔语中仍然是一个未触及的话题。为此,我们提出了尼泊尔语文本的抽象摘要。在这里,我们首先通过从在线新闻门户网站上抓取尼泊尔新闻来创建一个尼泊尔文文本数据集。其次,我们设计了一个基于深度学习的文本摘要模型,该模型基于带注意的编码器-解码器递归神经网络。更准确地说,长短期记忆(LSTM)单元用于编码器和解码器层。第三,我们通过选择不同的超参数(如隐藏层数和节点数)构建了9个不同的模型。最后,我们报告了每个模型的面向回忆的注册评价(ROUGE)分数,以评估它们的性能。在调整不同层数和隐藏状态创建的9个不同模型中,采用单层编码器和256个隐藏状态的模型在ROUGE-1 ROUGE-2和ROUGE-L的F-Score值分别为15.74、3.29和15.21,优于其他所有模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信