Complex query answering over knowledge graphs foundation model using region embeddings on a lie group

Zhengyun Zhou, Guojia Wan, Shirui Pan, Jia Wu, Wenbin Hu, Bo Du
{"title":"Complex query answering over knowledge graphs foundation model using region embeddings on a lie group","authors":"Zhengyun Zhou, Guojia Wan, Shirui Pan, Jia Wu, Wenbin Hu, Bo Du","doi":"10.1007/s11280-024-01254-7","DOIUrl":null,"url":null,"abstract":"<p>Answering complex queries with First-order logical operators over knowledge graphs, such as conjunction (<span>\\(\\wedge \\)</span>), disjunction (<span>\\(\\vee \\)</span>), and negation (<span>\\(\\lnot \\)</span>) is immensely useful for identifying missing knowledge. Recently, neural symbolic reasoning methods have been proposed to map entities and relations into a continuous real vector space and model logical operators as differential neural networks. However, traditional methodss employ negative sampling, which corrupts complex queries to train embeddings. Consequently, these embeddings are susceptible to divergence in the open manifold of <span>\\(\\mathbb {R}^n\\)</span>. The appropriate regularization is crucial for addressing the divergence of embeddings. In this paper, we introduces a Lie group as a compact embedding space for complex query embedding, enhancing ability to handle the intricacies of knowledge graphs the foundation model. Our method aims to solve the query of disjunctive and conjunctive problems. Entities and queries are represented as a region of a high-dimensional torus, where the projection, intersection, union, and negation of the torus naturally simulate entities and queries. After simulating the operations on the region of the torus we defined, we found that the resulting geometry remains unchanged. Experiments show that our method achieved a significant improvement on FB15K, FB15K-237, and NELL995. Through extensive experiments on datasets FB15K, FB15K-237, and NELL995, our approach demonstrates significant improvements, leveraging the strengths of knowledge graphs foundation model and complex query processing.</p>","PeriodicalId":501180,"journal":{"name":"World Wide Web","volume":"26 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"World Wide Web","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s11280-024-01254-7","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Answering complex queries with First-order logical operators over knowledge graphs, such as conjunction (\(\wedge \)), disjunction (\(\vee \)), and negation (\(\lnot \)) is immensely useful for identifying missing knowledge. Recently, neural symbolic reasoning methods have been proposed to map entities and relations into a continuous real vector space and model logical operators as differential neural networks. However, traditional methodss employ negative sampling, which corrupts complex queries to train embeddings. Consequently, these embeddings are susceptible to divergence in the open manifold of \(\mathbb {R}^n\). The appropriate regularization is crucial for addressing the divergence of embeddings. In this paper, we introduces a Lie group as a compact embedding space for complex query embedding, enhancing ability to handle the intricacies of knowledge graphs the foundation model. Our method aims to solve the query of disjunctive and conjunctive problems. Entities and queries are represented as a region of a high-dimensional torus, where the projection, intersection, union, and negation of the torus naturally simulate entities and queries. After simulating the operations on the region of the torus we defined, we found that the resulting geometry remains unchanged. Experiments show that our method achieved a significant improvement on FB15K, FB15K-237, and NELL995. Through extensive experiments on datasets FB15K, FB15K-237, and NELL995, our approach demonstrates significant improvements, leveraging the strengths of knowledge graphs foundation model and complex query processing.

Abstract Image

使用谎言群上的区域嵌入对知识图谱基础模型进行复杂查询回答
用知识图谱上的一阶逻辑运算符,如合(conjunction)、析(disjunction)和否(negative)来回答复杂的查询,对于识别缺失的知识非常有用。最近,有人提出了神经符号推理方法,将实体和关系映射到连续的实向量空间,并将逻辑运算符建模为微分神经网络。然而,传统方法ss 采用负采样,破坏了训练嵌入的复杂查询。因此,这些嵌入容易在开放流形(\mathbb {R}^n\ )中发生发散。适当的正则化对于解决嵌入发散问题至关重要。在本文中,我们引入了李群作为复杂查询嵌入的紧凑嵌入空间,增强了处理复杂知识图谱基础模型的能力。我们的方法旨在解决断条件和连接问题的查询。实体和查询被表示为高维环的一个区域,环的投影、交集、联合和否定自然地模拟实体和查询。在对我们定义的环形区域进行模拟操作后,我们发现所得到的几何图形保持不变。实验表明,我们的方法在 FB15K、FB15K-237 和 NELL995 上取得了显著的改进。通过在数据集 FB15K、FB15K-237 和 NELL995 上的广泛实验,我们的方法利用知识图基础模型和复杂查询处理的优势,取得了显著的改进。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信