Hi, I have to make an application in which,
The user asks a question, Google's API is used to convert the speech to text. But the problem is due to different accent the translator misunderstands the words. I want my application to guess the word to the nearest word spoken by the user. When the user speaks a word, the syllables in the word should be matched with the syllables of other words in my dictionary. Till now what I have done, I used CMU pronouncing dictionary. It contains ~122300 words along with its syllables. Along with this, I have created my own dictionary which consists of words that I need only. I have implemented a python logic that makes reasonable guesses for the words user speaks and matches with my dictionary. But sometimes for more than a word it makes the same guess. example: The user speaks "Light". The system translates it as "Bright" The user speaks "White" The system translates it as "Bright" I don't want this to happen. This system should make a reasonable conversion. Any help would be appreciated. I am using python for logic. Regards, Naveen BM -- https://mail.python.org/mailman/listinfo/python-list