Meng Gao , Yutao Xie , Wei Chen , Feng Zhang , Fei Ding , Tengjiao Wang , Jiahui Yao , Jiabin Zheng , Kam-Fai Wong
{"title":"多模态知识图谱补全的协同检索和重新排序框架","authors":"Meng Gao , Yutao Xie , Wei Chen , Feng Zhang , Fei Ding , Tengjiao Wang , Jiahui Yao , Jiabin Zheng , Kam-Fai Wong","doi":"10.1016/j.neunet.2025.107467","DOIUrl":null,"url":null,"abstract":"<div><div>Multi-modal knowledge graph completion (MMKGC) aims to predict missing links using entity’s multi-modal attributes. Embedding-based methods excel in leveraging structural knowledge, making them robust to entity ambiguity, yet their performance is constrained by the underutilization of multi-modal knowledge. Conversely, fine-tune-based (FT-based) approaches excel in extracting multi-modal knowledge but are hindered by ambiguity issues. To harness the complementary strengths of both methods for MMKGC, this paper introduces an ensemble framework <em>ReranKGC</em>, which decomposes KGC to a retrieve-and-rerank pipeline. The retriever employs embedding-based methods for initial retrieval. The re-ranker adopts our proposed KGC-CLIP, an FT-based method that utilizes CLIP to extract multi-modal knowledge from attributes for candidate re-ranking. By leveraging a more comprehensive knowledge source, the retriever generates a candidate pool containing entities not only semantically, but also structurally related to the query entity. Within this higher-quality candidate pool, the re-ranker can better discern candidates’ semantics to further refine the initial ranking, thereby enhancing precision. Through cooperation, each method maximizes its strengths while mitigating the weaknesses of others to a certain extent, leading to superior performance that surpasses individual capabilities. Extensive experiments conducted on link prediction tasks demonstrate that our framework ReranKGC consistently enhances baseline performance, outperforming state-of-the-art models.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"188 ","pages":"Article 107467"},"PeriodicalIF":6.0000,"publicationDate":"2025-04-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ReranKGC: A cooperative retrieve-and-rerank framework for multi-modal knowledge graph completion\",\"authors\":\"Meng Gao , Yutao Xie , Wei Chen , Feng Zhang , Fei Ding , Tengjiao Wang , Jiahui Yao , Jiabin Zheng , Kam-Fai Wong\",\"doi\":\"10.1016/j.neunet.2025.107467\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Multi-modal knowledge graph completion (MMKGC) aims to predict missing links using entity’s multi-modal attributes. Embedding-based methods excel in leveraging structural knowledge, making them robust to entity ambiguity, yet their performance is constrained by the underutilization of multi-modal knowledge. Conversely, fine-tune-based (FT-based) approaches excel in extracting multi-modal knowledge but are hindered by ambiguity issues. To harness the complementary strengths of both methods for MMKGC, this paper introduces an ensemble framework <em>ReranKGC</em>, which decomposes KGC to a retrieve-and-rerank pipeline. The retriever employs embedding-based methods for initial retrieval. The re-ranker adopts our proposed KGC-CLIP, an FT-based method that utilizes CLIP to extract multi-modal knowledge from attributes for candidate re-ranking. By leveraging a more comprehensive knowledge source, the retriever generates a candidate pool containing entities not only semantically, but also structurally related to the query entity. Within this higher-quality candidate pool, the re-ranker can better discern candidates’ semantics to further refine the initial ranking, thereby enhancing precision. Through cooperation, each method maximizes its strengths while mitigating the weaknesses of others to a certain extent, leading to superior performance that surpasses individual capabilities. Extensive experiments conducted on link prediction tasks demonstrate that our framework ReranKGC consistently enhances baseline performance, outperforming state-of-the-art models.</div></div>\",\"PeriodicalId\":49763,\"journal\":{\"name\":\"Neural Networks\",\"volume\":\"188 \",\"pages\":\"Article 107467\"},\"PeriodicalIF\":6.0000,\"publicationDate\":\"2025-04-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neural Networks\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0893608025003466\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0893608025003466","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
ReranKGC: A cooperative retrieve-and-rerank framework for multi-modal knowledge graph completion
Multi-modal knowledge graph completion (MMKGC) aims to predict missing links using entity’s multi-modal attributes. Embedding-based methods excel in leveraging structural knowledge, making them robust to entity ambiguity, yet their performance is constrained by the underutilization of multi-modal knowledge. Conversely, fine-tune-based (FT-based) approaches excel in extracting multi-modal knowledge but are hindered by ambiguity issues. To harness the complementary strengths of both methods for MMKGC, this paper introduces an ensemble framework ReranKGC, which decomposes KGC to a retrieve-and-rerank pipeline. The retriever employs embedding-based methods for initial retrieval. The re-ranker adopts our proposed KGC-CLIP, an FT-based method that utilizes CLIP to extract multi-modal knowledge from attributes for candidate re-ranking. By leveraging a more comprehensive knowledge source, the retriever generates a candidate pool containing entities not only semantically, but also structurally related to the query entity. Within this higher-quality candidate pool, the re-ranker can better discern candidates’ semantics to further refine the initial ranking, thereby enhancing precision. Through cooperation, each method maximizes its strengths while mitigating the weaknesses of others to a certain extent, leading to superior performance that surpasses individual capabilities. Extensive experiments conducted on link prediction tasks demonstrate that our framework ReranKGC consistently enhances baseline performance, outperforming state-of-the-art models.
期刊介绍:
Neural Networks is a platform that aims to foster an international community of scholars and practitioners interested in neural networks, deep learning, and other approaches to artificial intelligence and machine learning. Our journal invites submissions covering various aspects of neural networks research, from computational neuroscience and cognitive modeling to mathematical analyses and engineering applications. By providing a forum for interdisciplinary discussions between biology and technology, we aim to encourage the development of biologically-inspired artificial intelligence.