IRMA-International.org: Creator of Knowledge
Information Resources Management Association
Advancing the Concepts & Practices of Information Resources Management in Modern Organizations

Comparison of Tied-Mixture and State-Clustered HMMs with Respect to Recognition Performance and Training Method

Comparison of Tied-Mixture and State-Clustered HMMs with Respect to Recognition Performance and Training Method
View Sample PDF
Author(s): Hiroyuki Segi (NHK (Japan Broadcasting Corporation) Science & Technology Research Laboratories, Japan), Kazuo Onoe (NHK (Japan Broadcasting Corporation), Science & Technology Research Laboratories, Japan), Shoei Sato (NHK (Japan Broadcasting Corporation) Science & Technology Research Laboratories, Japan), Akio Kobayashi (NHK (Japan Broadcasting Corporation) Science & Technology Research Laboratories, Japan)and Akio Ando (University of Toyoma, Japan)
Copyright: 2017
Pages: 17
Source title: Artificial Intelligence: Concepts, Methodologies, Tools, and Applications
Source Author(s)/Editor(s): Information Resources Management Association (USA)
DOI: 10.4018/978-1-5225-1759-7.ch082

Purchase


Abstract

Tied-mixture HMMs have been proposed as the acoustic model for large-vocabulary continuous speech recognition and have yielded promising results. They share base-distribution and provide more flexibility in choosing the degree of tying than state-clustered HMMs. However, it is unclear which acoustic models to superior to the other under the same training data. Moreover, LBG algorithm and EM algorithm, which are the usual training methods for HMMs, have not been compared. Therefore in this paper, the recognition performance of the respective HMMs and the respective training methods are compared under the same condition. It was found that the number of parameters and the word error rate for both HMMs are equivalent when the number of codebooks is sufficiently large. It was also found that training method using the LBG algorithm achieves a 90% reduction in training time compared to training method using the EM algorithm, without degradation of recognition accuracy.

Related Content

Kamel Mouloudj, Vu Lan Oanh LE, Achouak Bouarar, Ahmed Chemseddine Bouarar, Dachel Martínez Asanza, Mayuri Srivastava. © 2024. 20 pages.
José Eduardo Aleixo, José Luís Reis, Sandrina Francisca Teixeira, Ana Pinto de Lima. © 2024. 52 pages.
Jorge Figueiredo, Isabel Oliveira, Sérgio Silva, Margarida Pocinho, António Cardoso, Manuel Pereira. © 2024. 24 pages.
Fatih Pinarbasi. © 2024. 20 pages.
Stavros Kaperonis. © 2024. 25 pages.
Thomas Rui Mendes, Ana Cristina Antunes. © 2024. 24 pages.
Nuno Geada. © 2024. 12 pages.
Body Bottom