Dear all:
        We are developing the multitouch framework for android system. Our
assumption is not to modify the android framework as far as possible.
So we will modify the multitouch driver and provide a
MultitouchAdapter.java to let the user program can listen some events,
like Zoom and Rotate! And the user program also can receive the
multitouch pointers to recogine the other gestures.

        The other work is to let the existing applicatons can support
multitouch gesture, like zoom event, without compling again. The
approach is to translate the Pointer Event to the Keyboard Event at
the driver layer because the user level program can't inject the
KeyEvent to other application due to security. For example, in the
synaptics driver, when detecting the zoom gesture, it will send the
KeyEvent to the keyboard device driver (/dev/input/event1). After the
browser receive the Menu+I Key combination, it will zoom in.

        Could anybody give us some suggestion? If our work is complete,
anyone needs it?
        We are afraid that the android team do the same thing now. Does
anyone know whether the official android team add the multitouch
support to android framework?

--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"Android Discuss" group.
To post to this group, send email to [email protected]
To unsubscribe from this group, send email to 
[email protected]
For more options, visit this group at 
http://groups.google.com/group/android-discuss?hl=en
-~----------~----~----~----~------~----~------~--~---

Reply via email to