论文标题
N-Gram语言模型的有效MDI改编
Efficient MDI Adaptation for n-gram Language Models
论文作者
论文摘要
本文提出了一种在最小歧视信息(MDI)原理下适应N-Gram语言模型适应的有效算法,其中对室外语言模型进行了调整以满足内域数据边际概率的约束。 MDI语言模型适应的挑战是其计算复杂性。通过利用N-Gram模型的向后结构以及最初提出的最大熵(ME)语言模型的层次训练方法的概念,我们表明可以以线性时间复杂性计算到每次迭代中输入的线性复杂性。尽管MDI比我更笼统,但复杂性与我的模型保持不变。这使MDI适应大型语料库和词汇。实验结果证实了我们在非常大的数据集上的算法的可伸缩性,而MDI适应性的困惑稍差,但与简单的线性插值相比,单词错误率结果更好。
This paper presents an efficient algorithm for n-gram language model adaptation under the minimum discrimination information (MDI) principle, where an out-of-domain language model is adapted to satisfy the constraints of marginal probabilities of the in-domain data. The challenge for MDI language model adaptation is its computational complexity. By taking advantage of the backoff structure of n-gram model and the idea of hierarchical training method, originally proposed for maximum entropy (ME) language models, we show that MDI adaptation can be computed in linear-time complexity to the inputs in each iteration. The complexity remains the same as ME models, although MDI is more general than ME. This makes MDI adaptation practical for large corpus and vocabulary. Experimental results confirm the scalability of our algorithm on very large datasets, while MDI adaptation gets slightly worse perplexity but better word error rate results compared to simple linear interpolation.