Shuai Yu , Wei Gao , Yongbin Qin , Caiwei Yang , Ruizhang Huang , Yanping Chen , Chuan Lin
{"title":"IterSum: Iterative summarization based on document topological structure","authors":"Shuai Yu , Wei Gao , Yongbin Qin , Caiwei Yang , Ruizhang Huang , Yanping Chen , Chuan Lin","doi":"10.1016/j.ipm.2024.103918","DOIUrl":null,"url":null,"abstract":"<div><div>Document structure plays a crucial role in understanding and analyzing document information. However, effectively encoding document structural features into the Transformer architecture faces significant challenges. This is primarily because different types of documents require the model to adopt varying structural encoding strategies, leading to a lack of a unified framework that can broadly adapt to different document types to leverage their structural properties. Despite the diversity of document types, sentences within a document are interconnected through semantic relationships, forming a topological semantic network. This topological structure is essential for integrating and summarizing information within the document. In this work, we introduce IterSum, a versatile text summarization framework applicable to various types of text. In IterSum, we utilize the document’s topological structure to divide the text into multiple blocks, first generating a summary for the initial block, then combining the current summary with the content of the next block to produce the subsequent summary, and continuing in this iterative manner until the final summary is generated. We validated our model on nine different types of public datasets, including news, knowledge bases, legal documents, and guidelines. Both quantitative and qualitative analyses were conducted, and the experimental results show that our model achieves state-of-the-art performance on all nine datasets measured by ROUGE scores. We also explored low-resource summarization, finding that even with only 10 or 100 samples in multiple datasets, top-notch results were obtained. Finally, we conducted human evaluations to further validate the superiority of our model.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":null,"pages":null},"PeriodicalIF":7.4000,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Processing & Management","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306457324002772","RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Document structure plays a crucial role in understanding and analyzing document information. However, effectively encoding document structural features into the Transformer architecture faces significant challenges. This is primarily because different types of documents require the model to adopt varying structural encoding strategies, leading to a lack of a unified framework that can broadly adapt to different document types to leverage their structural properties. Despite the diversity of document types, sentences within a document are interconnected through semantic relationships, forming a topological semantic network. This topological structure is essential for integrating and summarizing information within the document. In this work, we introduce IterSum, a versatile text summarization framework applicable to various types of text. In IterSum, we utilize the document’s topological structure to divide the text into multiple blocks, first generating a summary for the initial block, then combining the current summary with the content of the next block to produce the subsequent summary, and continuing in this iterative manner until the final summary is generated. We validated our model on nine different types of public datasets, including news, knowledge bases, legal documents, and guidelines. Both quantitative and qualitative analyses were conducted, and the experimental results show that our model achieves state-of-the-art performance on all nine datasets measured by ROUGE scores. We also explored low-resource summarization, finding that even with only 10 or 100 samples in multiple datasets, top-notch results were obtained. Finally, we conducted human evaluations to further validate the superiority of our model.
期刊介绍:
Information Processing and Management is dedicated to publishing cutting-edge original research at the convergence of computing and information science. Our scope encompasses theory, methods, and applications across various domains, including advertising, business, health, information science, information technology marketing, and social computing.
We aim to cater to the interests of both primary researchers and practitioners by offering an effective platform for the timely dissemination of advanced and topical issues in this interdisciplinary field. The journal places particular emphasis on original research articles, research survey articles, research method articles, and articles addressing critical applications of research. Join us in advancing knowledge and innovation at the intersection of computing and information science.