Skip to main navigation Skip to search Skip to main content

EFFICIENT COMPUTATION OF MMI NEURAL NETWORKS FOR LARGE VOCABULARY SPEECH RECOGNITION SYSTEMS

Research output: Contribution to conferencePaperpeer-review

Abstract

This paper describes, how to train Maximum Mutual Information Neural Networks (MMINN) in an efficient way, with a new topology. Large vocabulary speech recognition systems, based on a Hybrid MMI/connectionist HMM combination, have shown good performance on several tasks [1] and [2]. MMINNs are trained to maximize the mutual information between the index of the winning output neuron (Winner-Takes-All network) and the phonetical class of the corresponding acoustic frame. One major problem of MMI-neural networks is the high computational effort, which is needed for the training of the neural networks. The computational effort is proportional to the input and output size of the neural network and to the number of training samples. This paper shows two approaches, that demonstrate, how these long training times can be reduced with very low or even no loss in recognition accuracy. This is achieved by the use of phonetical knowledge, to build a network topology based on phonetical classes.

Original languageEnglish
StatePublished - 1998
Externally publishedYes
Event5th International Conference on Spoken Language Processing, ICSLP 1998 - Sydney, Australia
Duration: 30 Nov 19984 Dec 1998

Conference

Conference5th International Conference on Spoken Language Processing, ICSLP 1998
Country/TerritoryAustralia
CitySydney
Period30/11/984/12/98

Fingerprint

Dive into the research topics of 'EFFICIENT COMPUTATION OF MMI NEURAL NETWORKS FOR LARGE VOCABULARY SPEECH RECOGNITION SYSTEMS'. Together they form a unique fingerprint.

Cite this