论文标题

通过常识性知识提取和注射词汇约束的文本生成

Lexically-constrained Text Generation through Commonsense Knowledge Extraction and Injection

论文作者

Li, Yikang, Goel, Pulkit, Rajendra, Varsha Kuppur, Singh, Har Simrat, Francis, Jonathan, Ma, Kaixin, Nyberg, Eric, Oltramari, Alessandro

论文摘要

有条件的文本生成一直是一项具有挑战性的任务,尚未从最新模型中看到人类水平的绩效。在这项工作中,我们专门针对公共基准,其中的目的是为给定的一组输入概念生成一个合理的句子。尽管在其他任务方面取得了进步,但在该数据集上进行了微调的大型预训练的语言模型通常会产生句法正确但在质上偏离人类对常识的句子。此外,生成的序列无法满足诸如匹配言论和完整概念覆盖的词汇要求。在本文中,我们探讨了常识性知识图如何在常识性推理和词汇约束的解码方面如何增强模型性能。我们提出了增强生成文本的语义正确性的策略,我们通过:从概念网络中提取常识关系,通过注意机制将这些关系注入统一的语言模型(UNILM),并通过输出约束执行上述词汇要求。通过进行几次消融,我们发现常识性注入可以产生与人类理解更加一致的句子,同时仍然符合词汇要求。

Conditional text generation has been a challenging task that is yet to see human-level performance from state-of-the-art models. In this work, we specifically focus on the Commongen benchmark, wherein the aim is to generate a plausible sentence for a given set of input concepts. Despite advances in other tasks, large pre-trained language models that are fine-tuned on this dataset often produce sentences that are syntactically correct but qualitatively deviate from a human understanding of common sense. Furthermore, generated sequences are unable to fulfill such lexical requirements as matching part-of-speech and full concept coverage. In this paper, we explore how commonsense knowledge graphs can enhance model performance, with respect to commonsense reasoning and lexically-constrained decoding. We propose strategies for enhancing the semantic correctness of the generated text, which we accomplish through: extracting commonsense relations from Conceptnet, injecting these relations into the Unified Language Model (UniLM) through attention mechanisms, and enforcing the aforementioned lexical requirements through output constraints. By performing several ablations, we find that commonsense injection enables the generation of sentences that are more aligned with human understanding, while remaining compliant with lexical requirements.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源