Re: [Python-mode] more speech driven how twos
Am 18.06.2011 15:00, schrieb Eric S. Johansson: BTW what are suitable returns from Emacs report functions for you. As choices coming into my mind for the moment are: - simple get it returned yes as in position = Emacs_fetch ("current character position", buffer = "focus") buffer_list = Emacs_fetch ("buffer list") that looks like a list of values as parse-partial-sexp returns for example, right? [ ... ] I didn't get here overnight. It took me 18 years to become injured and 10 years to become frustrated with speech recognition and then another five years to figure out how to not become frustrated only become frustrated because I couldn't pay people to write the code for me. joys of being a serial entrepreneur self-employed type person. If you think this is interesting, you should see what I'm doing for diabetes self-management tools. I really need to get that done so I can pass the demo to potential funders which is part of the reason why I need these Emacs extensions. Everybody has an ulterior motive. :-) Even if neither Emacs nor Python related, I'll take occasion to mention Chaumont Devin, a poet from Hawai, who wrote some interesting computer linguistics and told a similar story beside. Here the adress of his language works: http://panlingua.net/ Cheers, Andreas ___ Python-mode mailing list Python-mode@python.org http://mail.python.org/mailman/listinfo/python-mode
Re: [Python-mode] more speech driven how twos
On 6/18/2011 9:00 AM, Eric S. Johansson wrote: I should probably into here because I'm probably giving you intellectual heartburn at this point because it's a lot to digest and I'm late for my astronomy club's work party. observing sites don't maintain themselves. one more bit of IHB. I believe that the model we are working with is leading us to a general foundation by which you can make all applications more easily accessible with greater functionality than the current efforts. I'm dead serious. The current accessibility methods have been make the same mistakes as long as I've been disabled and they provide no more accessibility today than they did in the past and vendor still have to do a lot to their applications to make it work. If we do this right, we make it easier, we make it simpler, and we make it possible for any vendor to make their software verifiably accessible without any great expense or effort. This is why I think this particular effort is important besides the minor point that it makes it possible for crips like me to use speech recognition and Emacs. it's kind of a big vision to take in but I think what I'm talking about will become visible as we make progress. --- eric ___ Python-mode mailing list Python-mode@python.org http://mail.python.org/mailman/listinfo/python-mode
Re: [Python-mode] more speech driven how twos
BTW what are suitable returns from Emacs report functions for you. As choices coming into my mind for the moment are: - simple get it returned yes as in position = Emacs_fetch ("current character position", buffer = "focus") buffer_list = Emacs_fetch ("buffer list") obviously we have to figure out what's a reasonable set of first arguments for this function. I am perfectly willing to use internal Lisp names including calling functions to get data. Actually, that seems like it might be very reasonable thing to do. Sort of like SQL queries only more understandable. - display in mode-line that seems to be more of a Emacs_set function. Could you elaborate what you're thinking of? - message-buffer actually, I was thinking of any buffer. If we had access to the current buffer, it would be possible to reimplement VR-mode in a more generic fashion. - tool-tip Again, could you elaborate but I do suspect it's something you would set through the set data function - so-called speed-bar Instead of a simple return it might be send so a program... My currently preferred Emacs is Xemacs for political reasons[1] I'm not sure what you need in a technical description. Normally in a speech recognition environment you use either fixed grammars or contiguous dictation. I am building a hybrid where you use a fixed grammar with contextually dependent elements and interact with GUI elements to make an unspeakable process speakable. the process of making the unspeakable speakable involves identifying and extracting information from the application and transforming it into a speakable form before displaying it in a second application which can be manipulated. See blog.esjworks.com for more complete examples. I expect that most of the action routines for a complete grammar will just be Emacs keystrokes invoking Emacs methods via keyboard input. It would be nice to do a direct injection of commands to eliminate problems with errors in command execution caused by too fast a rate of injecting characters. A direct access channel would also allows to query the buffer for state information which could be used to influence the action routine. The commands I asked for it which have no need to export information to any external program would help me get a better feel for if I'm on the right track or not. If there's something I use regularly and they "feel" right" is a vocal damage through excessive use, then I'm on the right path. If not, I need to look at the problem again they come up with a better solution. An example of a more complicated spoken command is the "get method" command. The first thing the command does is search to the right for the next method. An alias for it would be get next method. Going in the other direction would be get previous method. Once the method was identified, it would be placed in the region, mark on the left, point on the right. The action routine for the grammar would then invoke a GUI helper program to manipulate symbol names that pass the existing name along to it. The resulting change method would be returned via a different grammar and action routine, "use < transformation type>", and the result would be placed back into the buffer replacing what was in the region. Making any sense? It does. However, it's a new and vast matter for me. So let's proceed step by step and see how it goes. I didn't get here overnight. It took me 18 years to become injured and 10 years to become frustrated with speech recognition and then another five years to figure out how to not become frustrated only become frustrated because I couldn't pay people to write the code for me. joys of being a serial entrepreneur self-employed type person. If you think this is interesting, you should see what I'm doing for diabetes self-management tools. I really need to get that done so I can pass the demo to potential funders which is part of the reason why I need these Emacs extensions. Everybody has an ulterior motive. :-) Let's start with the report-function, telling where you are in code. Agreed? So I will dig a little bit into the question, how the results from Emacs are taken up in your environment. this has been problematic from day one. The blind spot in speech recognition has been that since getting information from application is "too hard" that commands have become open loop commands using the content of the command to generate a keystroke or menu injection commands to activate a function within an application. Emacs examples would be something like: (date | time | log) stamp = stamp.string($1); search (forward={ctrl+s}|back={ctrl+r}) = $1; yank (again={Esc}y|it back={ctrl+y}) = $1; go to line = {Alt+g}; repeat (it={ctrl+u} | twice={ctrl+u2} | thrice={ctrl+u3} ) = $1; copy(character={esc}xdelete-forward-char{enter}{ctrl+y} | word= {esc}d{ctrl+y}| line={esc}xset-mark{ctrl-e}{esc}w )= $1; kill(character = {ctrl+d} | word
Re: [Python-mode] more speech driven how twos
Am 17.06.2011 19:41, schrieb Eric S. Johansson: On 6/17/2011 12:18 PM, Andreas Röhler wrote: Am 17.06.2011 16:02, schrieb Eric S. Johansson: making more progress on some of the tools I need for speech or programming. [ ... ] Hi Eric, thanks reminding at that. Seeing you introduced a blueprint. Your previous text went into the `second-level-commands'. Yeah that was probably a speech recognition error. Speech and browsers do not get along well IIRC that's me who put it into some times ago :) [ ... ] use Dragon NaturallySpeaking for speech recognition system. Runs on Windows I use a community built extension called vocola to generate macros and some degree of smart user interface with Python extensions to vocola. Thanks, see it. BTW what are suitable returns from Emacs report functions for you. As choices coming into my mind for the moment are: - simple get it returned - display in mode-line - message-buffer - tool-tip - so-called speed-bar Instead of a simple return it might be send so a program... My currently preferred Emacs is Xemacs for political reasons[1] I'm not sure what you need in a technical description. Normally in a speech recognition environment you use either fixed grammars or contiguous dictation. I am building a hybrid where you use a fixed grammar with contextually dependent elements and interact with GUI elements to make an unspeakable process speakable. the process of making the unspeakable speakable involves identifying and extracting information from the application and transforming it into a speakable form before displaying it in a second application which can be manipulated. See blog.esjworks.com for more complete examples. I expect that most of the action routines for a complete grammar will just be Emacs keystrokes invoking Emacs methods via keyboard input. It would be nice to do a direct injection of commands to eliminate problems with errors in command execution caused by too fast a rate of injecting characters. A direct access channel would also allows to query the buffer for state information which could be used to influence the action routine. The commands I asked for it which have no need to export information to any external program would help me get a better feel for if I'm on the right track or not. If there's something I use regularly and they "feel" right" is a vocal damage through excessive use, then I'm on the right path. If not, I need to look at the problem again they come up with a better solution. An example of a more complicated spoken command is the "get method" command. The first thing the command does is search to the right for the next method. An alias for it would be get next method. Going in the other direction would be get previous method. Once the method was identified, it would be placed in the region, mark on the left, point on the right. The action routine for the grammar would then invoke a GUI helper program to manipulate symbol names that pass the existing name along to it. The resulting change method would be returned via a different grammar and action routine, "use < transformation type>", and the result would be placed back into the buffer replacing what was in the region. Making any sense? It does. However, it's a new and vast matter for me. So let's proceed step by step and see how it goes. Let's start with the report-function, telling where you are in code. Agreed? So I will dig a little bit into the question, how the results from Emacs are taken up in your environment. Cheers, Andreas ___ Python-mode mailing list Python-mode@python.org http://mail.python.org/mailman/listinfo/python-mode