Yingjie Song , Li Yang , Wenming Luo , Xiong Xiao , Zhuo Tang
{"title":"Boosting multi-document summarization with hierarchical graph convolutional networks","authors":"Yingjie Song , Li Yang , Wenming Luo , Xiong Xiao , Zhuo Tang","doi":"10.1016/j.neucom.2024.128753","DOIUrl":null,"url":null,"abstract":"<div><div>The input of the multi-document summarization task is usually long and has high redundancy. Encoding multiple documents is a challenge for the Seq2Seq architecture. The way of concatenating multiple documents into a sequence ignores the relation between documents. Attention-based Seq2Seq architectures have slightly improved the cross-document relation modeling for multi-document summarization. However, these methods ignore the relation between sentences, and there is little improvement that can be achieved through the attention mechanism alone. This paper proposes a hierarchical approach to leveraging the relation between words, sentences, and documents for abstractive multi-document summarization. Our model employs the Graph Convolutional Networks (GCN) for capturing the cross-document and cross-sentence relations. The GCN module can enrich semantic representations by generating high-level hidden features. Our model achieves significant improvement over the attention-based baseline, beating the Hierarchical Transformer by 3.4/1.64, 1.92/1.44 ROUGE-1/2 F1 points on the Multi-News and WikiSum datasets, respectively. Experimental results demonstrate that our delivered method brings substantial improvements over several strong baselines.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128753"},"PeriodicalIF":5.5000,"publicationDate":"2024-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231224015248","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The input of the multi-document summarization task is usually long and has high redundancy. Encoding multiple documents is a challenge for the Seq2Seq architecture. The way of concatenating multiple documents into a sequence ignores the relation between documents. Attention-based Seq2Seq architectures have slightly improved the cross-document relation modeling for multi-document summarization. However, these methods ignore the relation between sentences, and there is little improvement that can be achieved through the attention mechanism alone. This paper proposes a hierarchical approach to leveraging the relation between words, sentences, and documents for abstractive multi-document summarization. Our model employs the Graph Convolutional Networks (GCN) for capturing the cross-document and cross-sentence relations. The GCN module can enrich semantic representations by generating high-level hidden features. Our model achieves significant improvement over the attention-based baseline, beating the Hierarchical Transformer by 3.4/1.64, 1.92/1.44 ROUGE-1/2 F1 points on the Multi-News and WikiSum datasets, respectively. Experimental results demonstrate that our delivered method brings substantial improvements over several strong baselines.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.