论文标题
思想的反思:通过求解线性系统在语言模型中呈现数值推理
Reflection of Thought: Inversely Eliciting Numerical Reasoning in Language Models via Solving Linear Systems
论文作者
论文摘要
关于自然语言的数值推理一直是研究界的长期目标。但是,事实证明,尖端的语言模型很难可靠地将其概括为广泛的数字,尽管它们已经熟练地超过了共同数字和简单数字。在本文中,我们提出了一种新颖的方法,以使用简单的锚固数字来引起和利用预训练的语言模型中隐藏的数值推理知识。具体而言,我们首先利用简单的数字作为锚点来探测语言模型中隐式推断的算术表达式,然后明确地将表达式应用于复数上以获取相应的答案。为了成向引起算术表达式,我们将任务转换为分析解决的线性系统。几个数值推理基准的实验结果表明,我们的方法显着提高了现有LMS的数值推理能力。更重要的是,我们的方法是无训练的,并且只是在推理阶段工作,使其在所有零射击,很少射击和微调方案中都可以在各种语言模型(GPT-3,T5,BART等)中实现一致的性能优势。
Numerical reasoning over natural language has been a long-standing goal for the research community. However, cutting-edge language models have proven difficult to reliably generalize to a broad range of numbers, although they have shown proficiency in reasoning over common and simple numbers. In this paper, we propose a novel method to elicit and exploit the numerical reasoning knowledge hidden in pre-trained language models using simple anchor numbers. Concretely, we first leverage simple numbers as anchors to probe the implicitly inferred arithmetic expressions from language models, and then explicitly apply the expressions on complex numbers to get corresponding answers. To inversely elicit arithmetic expressions, we transform and formulate the task as an analytically solvable linear system. Experimental results on several numerical reasoning benchmarks demonstrate that our approach significantly improves numerical reasoning capabilities of existing LMs. More importantly, our approach is training-free and simply works in the inference phase, making it highly portable and achieving consistent performance benefits across a variety of language models (GPT-3, T5, BART, etc) in all zero-shot, few-shot, and fine-tuning scenarios.