Language Model Score Regularization for Speech Recognition

Inspired by the fact that back-off and interpolated smoothing algorithms have significant effect on statistical language modeling, this paper proposes a sentence-level Language model (LM) score regularization algorithm to improve the fault-tolerance of LMs for recognition errors. The proposed algori...

Full description

Saved in:
Bibliographic Details
Published inChinese Journal of Electronics Vol. 28; no. 3; pp. 604 - 609
Main Authors Zhang, Yike, Zhang, Pengyuan, Yan, Yonghong
Format Journal Article
LanguageEnglish
Published Published by the IET on behalf of the CIE 01.05.2019
Subjects
Online AccessGet full text
ISSN1022-4653
2075-5597
DOI10.1049/cje.2019.03.015

Cover

More Information
Summary:Inspired by the fact that back-off and interpolated smoothing algorithms have significant effect on statistical language modeling, this paper proposes a sentence-level Language model (LM) score regularization algorithm to improve the fault-tolerance of LMs for recognition errors. The proposed algorithm is applicable to both count-based LMs and neural network LMs. Instead of predicting the occurrence of a sequence of words under a fixed order Markov assumption, we use a composite model consisting of different order models with either n-gram or skip-gram features to estimate the probability of the sequence of words. In order to simplify implementations, we derive a connection between bidirectional neural networks and the proposed algorithm. Experiments were carried out on the Switchboard corpus. Results on N-best lists re-scoring show that the proposed algorithm achieves consistent word error rate reduction when it is applied to count-based LMs, Feedforward neural network (FNN) LMs, and Recurrent neural network (RNN) LMs.
ISSN:1022-4653
2075-5597
DOI:10.1049/cje.2019.03.015