Minimum classification error training of landmark models for real-time continuous speech recognition

Though many studies have shown the effectiveness of the minimum classification error (MCE) approach to discriminative training of HMM for speech recognition, few if any have reported MCE results for large (> 100 hours) training sets in the context of real-world, continuous speech recognition. Her...

Full description

Saved in:
Bibliographic Details
Published in2004 IEEE International Conference on Acoustics, Speech and Signal Processing Vol. 1; pp. I - 937
Main Authors McDermott, E., Hazen, T.J.
Format Conference Proceeding
LanguageEnglish
Japanese
Published Piscataway, N.J IEEE 28.09.2004
Subjects
Online AccessGet full text
ISBN9780780384842
0780384849
ISSN1520-6149
DOI10.1109/ICASSP.2004.1326141

Cover

More Information
Summary:Though many studies have shown the effectiveness of the minimum classification error (MCE) approach to discriminative training of HMM for speech recognition, few if any have reported MCE results for large (> 100 hours) training sets in the context of real-world, continuous speech recognition. Here we report large gains in performance for the MIT JUPITER weather information task as a result of MCE-based batch optimization of acoustic models. Investigation of word error rate versus computation time showed that small MCE models significantly outperform the maximum likelihood (ML) baseline at all points of equal computation time, resulting in up to 20% word error rate reduction for in-vocabulary utterances. The overall MCE loss function was minimized using Quickprop, a simple but effective second-order optimization method suited to parallelization over large training sets.
ISBN:9780780384842
0780384849
ISSN:1520-6149
DOI:10.1109/ICASSP.2004.1326141