TY - GEN
T1 - Scaled likelihood linear regression for hidden markov model adaptation
AU - Wallhoff, Frank
AU - Willetty, Daniel
AU - Rigoll, Gerhard
PY - 2001
Y1 - 2001
N2 - In the context of continuous Hidden Markov Model (HMM) based speech-recognition, linear regression approaches have become popular to adapt the acoustic models to the specific speaker's characteristics. The well known Maximum Likelihood Linear Regression (MLLR) [1] and Maximum A Posteriori Linear Regression (MAPLR) [2] are just two of them, which differ primarily in the training objective they are maximizing. However, besides the approaches mentioned above there exists another known training objective which is the Maximum Mutual Information (MMI). By combining this MMI-Approach with the linear regression of the HMM's mean values, our research group developed a new adaptation technique that we call Scaled Likelihood Linear Regression (SLLR) as introduced in [3]. In this approach, the distance of the correct model sequence against the wrong ones is discriminated framewise. Like all techniques using MMI objectives, this adaptation is computationally very expensive compared to techniques using ordinary ML based objectives. This paper therefore addresses the problem of an appropriate approximation technique to speed up this adaptation approach, by pruning the computation for tiny values in the discrimination objective. To further explore the potential of this adaptation technique and its approximation, the performance is measured on the LVCSR-system DUDeutsch developed by our research group at the Duisburg University and additionally on the 1993 WSJ adaptation tests of native and non-native speakers for the supervised case.
AB - In the context of continuous Hidden Markov Model (HMM) based speech-recognition, linear regression approaches have become popular to adapt the acoustic models to the specific speaker's characteristics. The well known Maximum Likelihood Linear Regression (MLLR) [1] and Maximum A Posteriori Linear Regression (MAPLR) [2] are just two of them, which differ primarily in the training objective they are maximizing. However, besides the approaches mentioned above there exists another known training objective which is the Maximum Mutual Information (MMI). By combining this MMI-Approach with the linear regression of the HMM's mean values, our research group developed a new adaptation technique that we call Scaled Likelihood Linear Regression (SLLR) as introduced in [3]. In this approach, the distance of the correct model sequence against the wrong ones is discriminated framewise. Like all techniques using MMI objectives, this adaptation is computationally very expensive compared to techniques using ordinary ML based objectives. This paper therefore addresses the problem of an appropriate approximation technique to speed up this adaptation approach, by pruning the computation for tiny values in the discrimination objective. To further explore the potential of this adaptation technique and its approximation, the performance is measured on the LVCSR-system DUDeutsch developed by our research group at the Duisburg University and additionally on the 1993 WSJ adaptation tests of native and non-native speakers for the supervised case.
UR - http://www.scopus.com/inward/record.url?scp=85009067171&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85009067171
T3 - EUROSPEECH 2001 - SCANDINAVIA - 7th European Conference on Speech Communication and Technology
SP - 1229
EP - 1232
BT - EUROSPEECH 2001 - SCANDINAVIA - 7th European Conference on Speech Communication and Technology
A2 - Lindberg, Borge
A2 - Benner, Henrik
A2 - Dalsgaard, Paul
A2 - Tan, Zheng-Hua
PB - International Speech Communication Association
T2 - 7th European Conference on Speech Communication and Technology - Scandinavia, EUROSPEECH 2001
Y2 - 3 September 2001 through 7 September 2001
ER -