On Wed, Jun 27, 2012 at 1:56 PM, Robert O'Callahan <[email protected]> wrote: > On Tue, Jun 26, 2012 at 8:22 AM, Tab Atkins Jr. <[email protected]>wrote: > >> The ability to capture sound and video from the user's devices and >> manipulate it in the page is already being exposed by the getUserMedia >> function. Theoretically, a Kinect can provide this information. >> >> More advanced functionality like Kinect's depth information probably >> needs more study and experience before we start thinking about adding >> it to the language itself. >> > > If we were going to support anything like this, I think the best approach > would be to have a new track type that getUserMedia can return in a > MediaStream, containing depth buffer data.
I agree. Experimentation with this in a non-live manner is already possible by using a @kind="metadata" track and putting the Kinect's depth information into a WebVTT file to use in parallel with the video. WebM has further defined how to encapsulate WebVTT into a WebM text track [1], so you could even put this information into a video file. I believe the same is possible with MPEG [2]. The exact format for how the Kinect's depth information is delivered as a timed metadata track would need to be specified before it could turn into its own @kind track type and deliver it live. Cheers, Silvia. [1] http://wiki.webmproject.org/webm-metadata/temporal-metadata/webvtt-in-webm [2] http://html5.cablelabs.com/tracks/media-container-mapping.html
