I am using Hmm for speech recognition of separate words. I have trained my Hmms for my database. I calculate and compare likelihood probabilities for an incoming audio signal. The problem I have is different words have different number of optimal states which will give different number of search paths (number of search paths = states^observations ) so probabilities can't be compared. How do I normalize the effect of different number of states?
Asked
Active
Viewed 401 times
1 Answers
4
You need either context free grammar or language model (usually - 3-gram probabilistic model) to recognize utterances rather than single words. Then you use appropriate algorithm to calculate score for each path. I strongly recommend you to take a look at existing solutions like Kaldi or CMUSphinx.

Alexander Solovets
- 2,447
- 15
- 22