{"title":"Attention-Seeker:无监督关键词提取的动态自我注意力评分","authors":"Erwin D. López Z., Cheng Tang, Atsushi Shimada","doi":"arxiv-2409.10907","DOIUrl":null,"url":null,"abstract":"This paper proposes Attention-Seeker, an unsupervised keyphrase extraction\nmethod that leverages self-attention maps from a Large Language Model to\nestimate the importance of candidate phrases. Our approach identifies specific\ncomponents - such as layers, heads, and attention vectors - where the model\npays significant attention to the key topics of the text. The attention weights\nprovided by these components are then used to score the candidate phrases.\nUnlike previous models that require manual tuning of parameters (e.g.,\nselection of heads, prompts, hyperparameters), Attention-Seeker dynamically\nadapts to the input text without any manual adjustments, enhancing its\npractical applicability. We evaluate Attention-Seeker on four publicly\navailable datasets: Inspec, SemEval2010, SemEval2017, and Krapivin. Our results\ndemonstrate that, even without parameter tuning, Attention-Seeker outperforms\nmost baseline models, achieving state-of-the-art performance on three out of\nfour datasets, particularly excelling in extracting keyphrases from long\ndocuments.","PeriodicalId":501281,"journal":{"name":"arXiv - CS - Information Retrieval","volume":"205 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Attention-Seeker: Dynamic Self-Attention Scoring for Unsupervised Keyphrase Extraction\",\"authors\":\"Erwin D. López Z., Cheng Tang, Atsushi Shimada\",\"doi\":\"arxiv-2409.10907\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper proposes Attention-Seeker, an unsupervised keyphrase extraction\\nmethod that leverages self-attention maps from a Large Language Model to\\nestimate the importance of candidate phrases. Our approach identifies specific\\ncomponents - such as layers, heads, and attention vectors - where the model\\npays significant attention to the key topics of the text. The attention weights\\nprovided by these components are then used to score the candidate phrases.\\nUnlike previous models that require manual tuning of parameters (e.g.,\\nselection of heads, prompts, hyperparameters), Attention-Seeker dynamically\\nadapts to the input text without any manual adjustments, enhancing its\\npractical applicability. We evaluate Attention-Seeker on four publicly\\navailable datasets: Inspec, SemEval2010, SemEval2017, and Krapivin. Our results\\ndemonstrate that, even without parameter tuning, Attention-Seeker outperforms\\nmost baseline models, achieving state-of-the-art performance on three out of\\nfour datasets, particularly excelling in extracting keyphrases from long\\ndocuments.\",\"PeriodicalId\":501281,\"journal\":{\"name\":\"arXiv - CS - Information Retrieval\",\"volume\":\"205 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Information Retrieval\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.10907\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Information Retrieval","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.10907","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Attention-Seeker: Dynamic Self-Attention Scoring for Unsupervised Keyphrase Extraction
This paper proposes Attention-Seeker, an unsupervised keyphrase extraction
method that leverages self-attention maps from a Large Language Model to
estimate the importance of candidate phrases. Our approach identifies specific
components - such as layers, heads, and attention vectors - where the model
pays significant attention to the key topics of the text. The attention weights
provided by these components are then used to score the candidate phrases.
Unlike previous models that require manual tuning of parameters (e.g.,
selection of heads, prompts, hyperparameters), Attention-Seeker dynamically
adapts to the input text without any manual adjustments, enhancing its
practical applicability. We evaluate Attention-Seeker on four publicly
available datasets: Inspec, SemEval2010, SemEval2017, and Krapivin. Our results
demonstrate that, even without parameter tuning, Attention-Seeker outperforms
most baseline models, achieving state-of-the-art performance on three out of
four datasets, particularly excelling in extracting keyphrases from long
documents.