Improved pronunciation modelling by inverse word frequency and pronunciation entropy
Tóm tắt
We propose a new approach to rank the potential pronunciations for each word by their pronunciation frequency and inverse word frequency (pf-iwf) weights. The pronunciation set obtained in this way can then be pruned with different criteria. This approach not only considers the frequencies of occurrence of the pronunciations, but tries to minimize the extra confusion which may be introduced by pronunciation variations, such that the best overall performance can be achieved. A new entropy-based approach for pruning the pronunciation variations is also proposed. Experimental results showed that the proposed approach can not only improve the recognition performance, but make the performance more stable and less sensitive to various parameters, factors and options including the different pruning criteria. All the experiments were performed with the LDC Mandarin Call Home corpus, although the approaches and principles are definitely not limited to Mandarin Chinese.
Từ khóa
#Inverse problems #Frequency #Entropy #Automatic speech recognition #Vocabulary #Natural languages #Costs #Training data #Dynamic programming #Heuristic algorithmsTài liệu tham khảo
wester, 2000, HA Comparison of Data-driven and Knowledge-based Modeling of Pronunciation variation, Proc ICSLP
williams, 1999, Knowing What You Don't Know: Roles for Confidence Measures in Automatic speech recognition
fosler, 1999, Dynamic Pronunciation Models for Automatic Speech Recognition
baeza, 1999, Modern Information Retrieval
fosler, 1999, Not just what, but also when: Guided Automatic Pronunciation Modeling for Broadcast News, DARPA Broadcast News Workshop
korkmazskiy, 1998, Statistical Modeling of Pronunciation and Production variations for Speech Recognition, Proc International Conference on Spoken Language Processing, 149
tsai, 2001, Pronunciation Variation Analysis with respect to Various Linguistic Levels and Contextual Conditions for Mandarin Chinese, Eurospeech
10.1016/S0167-6393(99)00036-9
10.1016/S0167-6393(99)00038-2
10.1016/S0167-6393(99)00037-0
