论文标题
RNN-TransDucers中隐性语言模型的自适应折现
Adaptive Discounting of Implicit Language Models in RNN-Transducers
论文作者
论文摘要
RNN-TransDucer(RNN-T)模型已成为流端到端ASR系统的代名词。尽管他们在许多评估类别上都具有竞争力,但稀有词对RNN-T模型构成了严重的挑战。稀有词表现降解的主要原因是,RNN-TS内部语言模型(LM)可能会过分自信,并导致幻觉预测与基本语音相关。为了解决此问题,我们建议使用轻巧的自适应LM折扣技术ADAPTLMD,可以与任何RNN-T体系结构一起使用,而无需任何外部资源或其他参数。 ADAPTLMD使用两种约束的方法:1)随机掩盖预测网络输出,以鼓励RNN-T不过分依赖其输出。 2)动态选择何时基于最近预测的令牌和ILM和隐式声学模型(IAM)分数之间的稀有性来打折隐式LM(ILM)。将ADAPTLMD与竞争性RNN-T基线进行比较,我们在对话式的,混合的印度英语ASR任务上,总体和稀有单词的总体和稀有单词分别相对减少了4%和14%。
RNN-Transducer (RNN-T) models have become synonymous with streaming end-to-end ASR systems. While they perform competitively on a number of evaluation categories, rare words pose a serious challenge to RNN-T models. One main reason for the degradation in performance on rare words is that the language model (LM) internal to RNN-Ts can become overconfident and lead to hallucinated predictions that are acoustically inconsistent with the underlying speech. To address this issue, we propose a lightweight adaptive LM discounting technique AdaptLMD, that can be used with any RNN-T architecture without requiring any external resources or additional parameters. AdaptLMD uses a two-pronged approach: 1) Randomly mask the prediction network output to encourage the RNN-T to not be overly reliant on it's outputs. 2) Dynamically choose when to discount the implicit LM (ILM) based on rarity of recently predicted tokens and divergence between ILM and implicit acoustic model (IAM) scores. Comparing AdaptLMD to a competitive RNN-T baseline, we obtain up to 4% and 14% relative reductions in overall WER and rare word PER, respectively, on a conversational, code-mixed Hindi-English ASR task.