论文标题

学习用关系抽象推理

Learning to Reason With Relational Abstractions

论文作者

Nam, Andrew J., Ren, Mengye, Finn, Chelsea, McClelland, James L.

论文摘要

大型语言模型最近在通过一系列解决方案步骤中进行了人体生成的序列进行微调时,显示了数学推理的有希望的进展。但是,解决方案序列不是正式结构化的,并且所得的模型生成的序列可能无法反映我们可能期望人类会产生的系统推理的种类。在本文中,我们研究了如何使用关系抽象的思想在语言模型中建立更强的推理能力。我们介绍了新的序列类型的序列,这些序列更明确地通过中间解决方案步骤向目标状态提供了对过渡的抽象表征。我们发现,以提示为提示的序列提供的模型可以以更高的精度解决任务,并且经过训练以产生此类序列的模型比接受过先前使用的人类生成的序列和其他基线的模型更好地解决了问题。因此,我们的工作采取了多个步骤,以阐明和改进语言模型在需要多步数学推理的任务上执行的方式。

Large language models have recently shown promising progress in mathematical reasoning when fine-tuned with human-generated sequences walking through a sequence of solution steps. However, the solution sequences are not formally structured and the resulting model-generated sequences may not reflect the kind of systematic reasoning we might expect an expert human to produce. In this paper, we study how to build stronger reasoning capability in language models using the idea of relational abstractions. We introduce new types of sequences that more explicitly provide an abstract characterization of the transitions through intermediate solution steps to the goal state. We find that models that are supplied with such sequences as prompts can solve tasks with a significantly higher accuracy, and models that are trained to produce such sequences solve problems better than those that are trained with previously used human-generated sequences and other baselines. Our work thus takes several steps toward elucidating and improving how language models perform on tasks requiring multi-step mathematical reasoning.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源