Combining Neural Networks And Hidden Markov Models For Continuous Speech Recognition
暂无分享,去创建一个
Pure MLP-based approaches have not previously been t demonstrated to function well for continuous-speech recogni ion because of the need for accurate segmentation of the w speech signal. HMMs, on the other hand, provide a frame ork for simultaneous segmentation and classification of t speech, which has been demonstrated to be useful for con inuous recognition. Previous work by Morgan and Bourlard d H [1] has shown theoretically and practically that MLPs an MMs can be combined by using MLPs for the estimation , t of the HMM state-dependent observation probabilities hereby exploiting the advantages of both approaches.
[1] Frederick Jelinek,et al. Interpolated estimation of Markov source parameters from sparse data , 1980 .
[2] John Makhoul,et al. Context-dependent modeling for acoustic-phonetic recognition of continuous speech , 1985, ICASSP '85. IEEE International Conference on Acoustics, Speech, and Signal Processing.
[3] Hervé Bourlard,et al. Continuous speech recognition using multilayer perceptrons with hidden Markov models , 1990, International Conference on Acoustics, Speech, and Signal Processing.