On Wed, Jan 9, 2013 at 11:49 AM, Arvind R <arvin...@gmail.com> wrote:
> Hi all, > > My understanding is that emotion gets the video backend to render RGBA > to the evas canvas that is then displayed by the ecore-evas backend. > Correct? > > If so, would it be possible, for instance, using the xine backend to > render directly to screen using whatever HW-accleration available to > it, and have the evas-canvas as an 'underlay' to the video screen in > order to trap events. This would mean modifying the emotion-xine > module to be an interceptor in the xine pipeline instead of being a > video_output driver. > > Feasible? > > Arvind R. > > > ------------------------------------------------------------------------------ > Master Java SE, Java EE, Eclipse, Spring, Hibernate, JavaScript, jQuery > and much more. Keep your Java skills current with LearnJavaNow - > 200+ hours of step-by-step video tutorials by Java experts. > SALE $49.99 this month only -- learn more at: > http://p.sf.net/sfu/learnmore_122612 > _______________________________________________ > enlightenment-users mailing list > enlightenment-users@lists.sourceforge.net > https://lists.sourceforge.net/lists/listinfo/enlightenment-users > -- Gustavo Sverzut Barbieri http://profusion.mobi embedded systems -------------------------------------- MSN: barbi...@gmail.com Skype: gsbarbieri Mobile: +55 (19) 9225-2202 ------------------------------------------------------------------------------ Master Java SE, Java EE, Eclipse, Spring, Hibernate, JavaScript, jQuery and much more. Keep your Java skills current with LearnJavaNow - 200+ hours of step-by-step video tutorials by Java experts. SALE $49.99 this month only -- learn more at: http://p.sf.net/sfu/learnmore_122612 _______________________________________________ enlightenment-users mailing list enlightenment-users@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/enlightenment-users