论文标题
洛伦:可解释的事实验证的逻辑登记的推理
LOREN: Logic-Regularized Reasoning for Interpretable Fact Verification
论文作者
论文摘要
鉴于自然语言陈述,如何验证其对Wikipedia(例如Wikipedia)的大规模文本知识来源的真实性?大多数现有的神经模型都会做出预测,而无需提供有关错误主张的哪一部分的线索。在本文中,我们提出了Loren,这是一种可解释的事实验证的方法。我们在短语级别上分解了整个主张的验证,该短语的真实性是解释的,可以根据逻辑规则将其汇总到最终判决中。洛伦的关键见解是将主张短语的真实性表示为三值潜在的潜在变量,这些变量是通过聚合逻辑规则正规化的。最终索赔验证基于所有潜在变量。因此,洛伦(Loren)享有可解释性的额外好处 - 很容易解释它如何以索赔短语真实性达到某些结果。公共事实验证基准的实验表明,洛伦在享有忠实和准确的解释性的功能的同时,与以前的方法具有竞争力。 Loren的资源可在以下网址获得:https://github.com/jiangjiechen/loren。
Given a natural language statement, how to verify its veracity against a large-scale textual knowledge source like Wikipedia? Most existing neural models make predictions without giving clues about which part of a false claim goes wrong. In this paper, we propose LOREN, an approach for interpretable fact verification. We decompose the verification of the whole claim at phrase-level, where the veracity of the phrases serves as explanations and can be aggregated into the final verdict according to logical rules. The key insight of LOREN is to represent claim phrase veracity as three-valued latent variables, which are regularized by aggregation logical rules. The final claim verification is based on all latent variables. Thus, LOREN enjoys the additional benefit of interpretability -- it is easy to explain how it reaches certain results with claim phrase veracity. Experiments on a public fact verification benchmark show that LOREN is competitive against previous approaches while enjoying the merit of faithful and accurate interpretability. The resources of LOREN are available at: https://github.com/jiangjiechen/LOREN.