{"title":"BERT-based Extractive Text Summarization of Scholarly Articles: A Novel Architecture","authors":"Sheher Bano, Shah Khalid","doi":"10.1109/ICAIoT57170.2022.10121826","DOIUrl":null,"url":null,"abstract":"Currently, there are a variety of extractive summarization approaches available, each with its own set of advantages and disadvantages. However, none of them are ideal, which means there is still room for advancement in this field of automation. BERT is a multilayer transformer network that has been pre-trained for a variety of self-supervised applications. However, because of its input length restriction, it is only appropriate for short text. As a result, we believe that using BERT for long document summarization will be a challenging task. We suggest a novel approach through which BERT can be utilized to summarize long documents. We used the method of dividing a whole document into multiple chunks and each chunk contains one sentence. The basic idea is to get sentence embeddings from BERT and then apply an encoder-decoder model on top of BERT. Experiments are conducted with two scholarly datasets (arXiv and PubMed). The results show that our technique consistently outperform several state-of-the-art models.","PeriodicalId":297735,"journal":{"name":"2022 International Conference on Artificial Intelligence of Things (ICAIoT)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Artificial Intelligence of Things (ICAIoT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAIoT57170.2022.10121826","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
Currently, there are a variety of extractive summarization approaches available, each with its own set of advantages and disadvantages. However, none of them are ideal, which means there is still room for advancement in this field of automation. BERT is a multilayer transformer network that has been pre-trained for a variety of self-supervised applications. However, because of its input length restriction, it is only appropriate for short text. As a result, we believe that using BERT for long document summarization will be a challenging task. We suggest a novel approach through which BERT can be utilized to summarize long documents. We used the method of dividing a whole document into multiple chunks and each chunk contains one sentence. The basic idea is to get sentence embeddings from BERT and then apply an encoder-decoder model on top of BERT. Experiments are conducted with two scholarly datasets (arXiv and PubMed). The results show that our technique consistently outperform several state-of-the-art models.