Player has several audio related interfaces:
-speech: implemented in the speech drivers that takes care of launching and killing festival along with the proper communication with it. -speech_recognition: The implementation depends on sphinx2 -audio: Currently seems disabled The fact is that those drivers are managed in a very model independent way by Player. So, I don't see any need to take those drivers and implement them in gazebo (say copy and paste). What I meant is that through Player it is perfectly possible to integrate text to speech and sound to a simulation. For instance, the wheels turns in Gazebo model and the client knows it thought the gazebo driver, then use the audio driver to play some funny sound. The only problem is that Gazebo has 2 interfaces: the one with player and the one named libgazebo. The second interface needs audio? In fact, we need that second interface? Why is Gazebo project maintaining libgazebo instead of modify wxgazebo to use player (similar to playerv)? I think that most people use Gazebo because it is used "the official" 3d sim for player, why don't accept this fact and stop maintaining libgazebo ? -- Jordi Polo ------------------------------------------------------------------------- Using Tomcat but need to do more? Need to support web services, security? Get stuff done quickly with pre-integrated technology to make your job easier Download IBM WebSphere Application Server v.1.0.1 based on Apache Geronimo http://sel.as-us.falkag.net/sel?cmd=lnk&kid=120709&bid=263057&dat=121642 _______________________________________________ Playerstage-gazebo mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/playerstage-gazebo
