论文标题
Unikgqa:统一的检索和解决多跳的问题,回答知识图
UniKGQA: Unified Retrieval and Reasoning for Solving Multi-hop Question Answering Over Knowledge Graph
论文作者
论文摘要
多跳的问题回答知识图〜(kgqa)旨在找到与大规模知识图(kg)上自然语言问题中提到的主题实体相远的答案实体。为了应对庞大的搜索空间,现有工作通常采用两阶段的方法:它首先检索与问题相关的相对较小的子图,然后在子图上执行推理以准确地找到答案实体。尽管这两个阶段高度相关,但以前的工作采用非常不同的技术解决方案来开发检索和推理模型,从而忽略了其在任务本质上的相关性。在本文中,我们提出了Unikgqa,这是一种通过在模型体系结构和参数学习中统一检索和推理来实现多跳KGQA任务的新方法。对于模型体系结构,UNIKGQA由一个基于预先训练的语言模型〜(PLM)的语义匹配模块组成,用于问题相关语义匹配,以及一个匹配的信息传播模块,以沿着KGS的有向边缘传播匹配的信息。对于参数学习,我们根据检索和推理模型的问题关联匹配设计共享的预训练任务,然后提出以检索和推理为导向的微调策略。与以前的研究相比,我们的方法更加统一,与检索和推理阶段紧密相关。在三个基准数据集上进行的广泛实验证明了我们方法对多跳KGQA任务的有效性。我们的代码和数据可在〜\ url {https://github.com/rucaibox/unikgqa}上公开获得。
Multi-hop Question Answering over Knowledge Graph~(KGQA) aims to find the answer entities that are multiple hops away from the topic entities mentioned in a natural language question on a large-scale Knowledge Graph (KG). To cope with the vast search space, existing work usually adopts a two-stage approach: it first retrieves a relatively small subgraph related to the question and then performs the reasoning on the subgraph to find the answer entities accurately. Although these two stages are highly related, previous work employs very different technical solutions for developing the retrieval and reasoning models, neglecting their relatedness in task essence. In this paper, we propose UniKGQA, a novel approach for multi-hop KGQA task, by unifying retrieval and reasoning in both model architecture and parameter learning. For model architecture, UniKGQA consists of a semantic matching module based on a pre-trained language model~(PLM) for question-relation semantic matching, and a matching information propagation module to propagate the matching information along the directed edges on KGs. For parameter learning, we design a shared pre-training task based on question-relation matching for both retrieval and reasoning models, and then propose retrieval- and reasoning-oriented fine-tuning strategies. Compared with previous studies, our approach is more unified, tightly relating the retrieval and reasoning stages. Extensive experiments on three benchmark datasets have demonstrated the effectiveness of our method on the multi-hop KGQA task. Our codes and data are publicly available at~\url{https://github.com/RUCAIBox/UniKGQA}.