The Mutual Information as a Scoring Function for Speech Recognition
スポンサーリンク
概要
- 論文の詳細を見る
In recent speech recognition technology, the score of a hypothesis is often defined on the basis of likelihood calculated with an HMM. As is well known, however, direct use of likelihood for score causes difficult problems especially in continuous speech recognition. In this work, the mutual information between a speech segment and a hypothesized word was employed as a scoring function, and the performance was tested from various points of view. The mutual information is obtained by normalizing the likelihood by a speech probability. In order to estimate the speech probability, an ergodic HMM was utilized. Through a number of experiments, it was confirmed that the mutual information was a significantly better scoring function than the log-likelihood. There is another well known normalization method, in which the likelihood is normalized by a speech probability estimated with an all-phone model. Comparison of the two normalization methods was also carried out, leading to a conclusion that the speech probability estimated with an ergodic HMM gave a better scoring function than that estimated with an all-phone model.
- 社団法人電子情報通信学会の論文
- 1995-12-15
著者
-
Ozeki Kazuhiko
Cstr The University Of Edinburgh:the University Of Electro-communications
-
Ozeki Kazuhiko
CSTR, The University of Edinburgh