Hi,

I have to make an application in which, 

The user asks a question, Google's API is used to convert the speech to text. 
But the problem is due to different accent the translator misunderstands the 
words. 
I want my application to guess the word to the nearest word spoken by the user. 
When the user speaks a word, the syllables in the word should be matched with 
the syllables of other words in my dictionary. 
Till now what I have done,
I used CMU pronouncing dictionary. It contains ~122300 words along with its 
syllables.
Along with this, I have created my own dictionary which consists of words that 
I need only.
I have implemented a python logic that makes reasonable guesses for the words 
user speaks and matches with my dictionary.
But sometimes for more than a word it makes the same guess.
example:
The user speaks "Light". The system translates it as "Bright"
The user speaks "White" The system translates it as "Bright"
I don't want this to happen. This system should make a reasonable conversion.

Any help would be appreciated.

I am using python for logic.

Regards,
Naveen BM
-- 
https://mail.python.org/mailman/listinfo/python-list

Reply via email to