Thanks Marcelo for the exaplanation. So *that's* why Siri sounds so good! I can see how articulatory synthesis, when fully developed, can be more powerful because you don't need to pre-record everything!
So is incorporating the research done on articulatory synthesis over the past decades the current development focus of Gnuspeech? What are the development focuses anyway? Also, as a non-programmer and complete non-expert on the subject, how can an user support and expedite development of Gnuspeech? Finally, other than Gnuspeech are there other Free Software text-to-speech software that can produce equal of better quality synthesis? Thanks!! On 01/11/2015, Marcelo Y. Matuda <[email protected]> wrote: > Hi, > > On 11/01/2015 03:45 PM, Advrk Aplmrkt wrote: >> Thanks for the links, and I agree a proper man page or quickstart >> guide would be super useful for end users! (and not just speech >> synthesis researchers) >> >> I checked out the YouTube videos, and I confess it was hard for me to >> understand what Gnuspeech was saying... Is there a reason why it >> doesn't sound nearly as natural as, say, Siri yet??? > > Siri uses a method called Unit Selection (AFAIK), which joins segments > of recorded speech. That is why the quality can be so good. > > Gnuspeech uses articulatory synthesis, which uses a mathematical model > of the human vocal tract to synthesize the speech from scratch. It is > very difficult to adjust the many parameters. Also GnuspeechSA is a C++ > port of the original TTS_Server (for NeXTSTEP), developed a long time > ago. It doesn't yet incorporate the research done in all these years. > Hopefully articulatory synthesis will reach the quality of unit > selection, but there is much work to do. > > Regards, > Marcelo _______________________________________________ gnuspeech-contact mailing list [email protected] https://lists.gnu.org/mailman/listinfo/gnuspeech-contact
