This was going to be my question during the dev summit: why *don't* we do this, and improve our detection over time (and retroactively) instead of relying on user's devices?
Which is a long way of saying: +1 :) -- Sent from my phone, please excuse brevity. On Feb 3, 2015 12:10 PM, "Max Semenik" <[email protected]> wrote: > Problem: while apps have face detection available from iOS and Android for > use in lead image positioning/cropping, mobile web doesn't have it, and > even for apps detection is quite slow, resulting in battery drain and > UX-problematic slowdown, especially on low-end Android devices. > > With Dmitry's help, I discovered Android's face detection library sources > ([1], separated out to a standalone library at [2]). This means that we can > build a face detection service and supply its results to all users, be that > apps, web or third parties. > > Thoughts? > > ---- > [1] https://android.googlesource.com/platform/external/neven/+/master > [2] https://github.com/lqs/neven > > -- > Best regards, > Max Semenik ([[User:MaxSem]]) > > _______________________________________________ > Mobile-l mailing list > [email protected] > https://lists.wikimedia.org/mailman/listinfo/mobile-l > >
_______________________________________________ Mobile-l mailing list [email protected] https://lists.wikimedia.org/mailman/listinfo/mobile-l
