高级检索

    基于生成模型的开放域关系推理方法

    Open Domain Relation Reasoning Based on Generative Model

    • 摘要: 关系推理是自然语言处理中的一项重要任务,旨在预测2个或多个实体之间可能存在的语义关系,推理过程通常为从已知的实体间关系中推导出新的关系,得到的结果可以在多种下游任务如知识图谱补全、关系抽取、常识知识问答中得到广泛应用. 以往的研究主要存在2个局限性:首先,以往的方法主要集中于封闭域,其中关系类型都是已经事先定义好的,难以扩展;其次,即便存在少量针对开放域的关系推理方法,也仅聚焦于单跳推理,难以满足更复杂的场景需求. 因此,定义了开放域的2跳关系推理任务,并构建了一个用于评估该任务的数据集. 面向该任务,提出了一种基于生成模型的开放域关系推理方法ORANGE,包括实体生成、关系生成、结果聚合3个模块. 实验结果表明,ORANGE相比现有主流关系推理方法在平均得分上提高了10.36%. 此外,当ORANGE的关系推理框架与大语言模型结合使用时,相较于传统的上下文学习提示策略,平均得分提高了9.58%.

       

      Abstract: Relation reasoning, an important task in natural language processing, aims to infer possible semantic relations between two or more entities. The reasoning process typically involves deriving new relations from known relations between entities, and the results can be widely applied in various downstream tasks such as knowledge graph completion, relation extraction, and commonsense knowledge question answering. Previous studies often face two main limitations: First, they are primarily based on the closed-world assumption, meaning the relation types are predefined and difficult to expand. Second, even if some methods focus on open domains, they typically only handle 1-hop reasoning, which is insufficient for complex multi-hop reasoning scenarios. To address these issues, we propose and define an open domain 2-hop relation reasoning task and construct a dataset for evaluating this task. Furthermore, we introduce an open domain 2-hop relation reasoning framework, named ORANGE (open domain relation reasoning method on generative model), which includes 3 key modules: entity generation, relation generation, and result aggregation. Firstly, the entity generation module generates unknown entities. Secondly, the relation generation module proposes potential new relations. Finally, the result aggregation module integrates the outputs of the preceding modules to determine the final result. Experimental results demonstrate that, when compared to the best existing methods, our approach achieves a 10.36% improvement in the average score. Moreover, when employing ORANGE’s 3-module relation reasoning framework with large language models, it surpasses the conventional in-context learning prompt strategy, showcasing a 9.58% enhancement in the average score.

       

    /

    返回文章
    返回