This was going to be my question during the dev summit: why *don't* we do
this, and improve our detection over time (and retroactively) instead of
relying on user's devices?

Which is a long way of saying: +1 :)

--
Sent from my phone, please excuse brevity.
On Feb 3, 2015 12:10 PM, "Max Semenik" <[email protected]> wrote:

> Problem: while apps have face detection available from iOS and Android for
> use in lead image positioning/cropping, mobile web doesn't have it, and
> even for apps detection is quite slow, resulting in battery drain and
> UX-problematic slowdown, especially on low-end Android devices.
>
> With Dmitry's help, I discovered Android's face detection library sources
> ([1], separated out to a standalone library at [2]). This means that we can
> build a face detection service and supply its results to all users, be that
> apps, web or third parties.
>
> Thoughts?
>
> ----
> [1] https://android.googlesource.com/platform/external/neven/+/master
> [2] https://github.com/lqs/neven
>
> --
> Best regards,
> Max Semenik ([[User:MaxSem]])
>
> _______________________________________________
> Mobile-l mailing list
> [email protected]
> https://lists.wikimedia.org/mailman/listinfo/mobile-l
>
>
_______________________________________________
Mobile-l mailing list
[email protected]
https://lists.wikimedia.org/mailman/listinfo/mobile-l

Reply via email to