论文标题
LED:词典启发的大规模检索的致密猎犬
LED: Lexicon-Enlightened Dense Retriever for Large-Scale Retrieval
论文作者
论文摘要
基于语义空间中密集表示的检索模型已成为第一阶段检索的必不可少的分支。这些检索员受益于表示朝着压缩全球序列层嵌入的表示的进步。但是,它们很容易忽略本地显着短语和实体在文本中提到的,这些短语通常在第一阶段的检索中扮演枢轴角色。为了减轻这种弱点,我们提议使一个密集的猎犬对齐表现出色的词典表示代表模型。通过弱化的知识蒸馏来实现对齐方式,以通过两个方面启发回猎犬 - 1)词汇扬声的对比目标,以挑战密集的编码器和2)一个配对的阵列矛盾的正规化,以使密集的模型的行为倾向于对方。我们在三个公共基准上评估了我们的模型,这表明,凭借可比的词典觉得猎犬作为老师,我们提议的密集人可以带来一致而重大的改进,甚至超过教师。此外,我们发现我们对浓犬的改进是与标准排名蒸馏的补充,这可以进一步提高最先进的性能。
Retrieval models based on dense representations in semantic space have become an indispensable branch for first-stage retrieval. These retrievers benefit from surging advances in representation learning towards compressive global sequence-level embeddings. However, they are prone to overlook local salient phrases and entity mentions in texts, which usually play pivot roles in first-stage retrieval. To mitigate this weakness, we propose to make a dense retriever align a well-performing lexicon-aware representation model. The alignment is achieved by weakened knowledge distillations to enlighten the retriever via two aspects -- 1) a lexicon-augmented contrastive objective to challenge the dense encoder and 2) a pair-wise rank-consistent regularization to make dense model's behavior incline to the other. We evaluate our model on three public benchmarks, which shows that with a comparable lexicon-aware retriever as the teacher, our proposed dense one can bring consistent and significant improvements, and even outdo its teacher. In addition, we found our improvement on the dense retriever is complementary to the standard ranker distillation, which can further lift state-of-the-art performance.