I think it requires kernel level monitor. Maybe you should start from getting source code of android. Write driver to enumarate all controls of foreground apps.
SJ On Mar 31, 4:57 am, Ella <daffodil...@gmail.com> wrote: > The same question from me. > > We know, A GUI automation testing tool -- monkey runner is there, it > provide some functions such as Keyboard function, input message > functions as well as touch, which are all very good supports for the > automation testing. > > However, there seems somethings that puzzle me a lot, so search helps > from all of you here: > 1. As in most of GUI automation testing, locate the objects in the > screen and judge the existing of the objects are very important, so is > there any way to implement this function, or can we extend the > function ourselves, if so, how? > 2. Now we can't touch the object by provide name or id or any other > attributes of the objects but just touch on location(by X,Y). in > common GUI automation, it's very inflexible and bring much trouble to > users to get the location info. Is there any way to resolve this > problems? or any further plans for the enhancement? > > Thanks! -- You received this message because you are subscribed to the Google Groups "Android Developers" group. To post to this group, send email to android-developers@googlegroups.com To unsubscribe from this group, send email to android-developers+unsubscr...@googlegroups.com For more options, visit this group at http://groups.google.com/group/android-developers?hl=en