论文标题

适应性:基于及时的NLP的自适应模型培训

AdaPrompt: Adaptive Model Training for Prompt-based NLP

论文作者

Chen, Yulong, Liu, Yang, Dong, Li, Wang, Shuohang, Zhu, Chenguang, Zeng, Michael, Zhang, Yue

论文摘要

基于迅速的学习能够解决零射击和几乎没有射击的NLP任务,在社区中引起了很多关注。主要思想是通过将这些任务映射到自然语言提示中,弥合NLP下游任务与语言建模(LM)之间的差距,然后将这些任务填充,然后由预训练的语言模型(PLM)填充。但是,对于迅速的学习,NLP任务和预处理之间仍然存在两个显着的差距。首先,在LM预处理期间,及时信息不一定足够。其次,特定于任务的数据不一定在审议过程中很好地表示。我们通过提出适应性来解决这两个问题,通过使用任务和及时特征来自适应地检索外部数据以持续预处理PLM。此外,我们利用自然语言推论模型中的知识来得出自适应的言语。五个NLP基准测试的实验结果表明,适应性可以在几个射击设置中改善标准PLM。此外,在零摄像设置中,我们的方法的表现优于基于标准的提示方法,最多可减少26.35 \%。

Prompt-based learning, with its capability to tackle zero-shot and few-shot NLP tasks, has gained much attention in community. The main idea is to bridge the gap between NLP downstream tasks and language modeling (LM), by mapping these tasks into natural language prompts, which are then filled by pre-trained language models (PLMs). However, for prompt learning, there are still two salient gaps between NLP tasks and pretraining. First, prompt information is not necessarily sufficiently present during LM pretraining. Second, task-specific data are not necessarily well represented during pretraining. We address these two issues by proposing AdaPrompt, adaptively retrieving external data for continual pretraining of PLMs by making use of both task and prompt characteristics. In addition, we make use of knowledge in Natural Language Inference models for deriving adaptive verbalizers. Experimental results on five NLP benchmarks show that AdaPrompt can improve over standard PLMs in few-shot settings. In addition, in zero-shot settings, our method outperforms standard prompt-based methods by up to 26.35\% relative error reduction.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源