Dear Olaf, This project is very interesting for us since most of the lecture recordings in TAU are made by students sitting behind the camera and tracing the lecturer. Though the stuff cost is not so high here we have quite a problem finding students to keep on working at least for a whole semester long:(. The most difficult problem in lecturer tracing system, in my opinion, is not the processes of tracing the lecturer or identify him from others crossing the camera frame but to make the intelligent decision whether to keep on tracking the lecturer while he moves away after writing something on the board or stop the tracking and remain on the written content on the blackboard previously written by the lecturer. I would like to know if there is an strategy on how to deal with this problem on the work of Benjamin Wulff ? Any way the demo clip is very impressive and I am sure it is a big step towards lecture recordings automation.
Cheers to Benjamin Wulff and others working with him on this project. Best Jack -----Original Message----- From: [email protected] [mailto:[email protected]] On Behalf Of Schulte Olaf A. Sent: Thursday, August 30, 2012 11:06 AM To: Opencast Matterhorn; Opencast Community; Opencast Matterhorn Subject: [Opencast] Automated camera tracking to be integrated with Opencast Matterhorn Dear All ELAN e.V. together with Osnabruck University and ETH Zurich would like to announce they are working towards an automated camera tracking system "LectureSight", fully integrated with Opencast Matterhorn. The module will provide a solution to automatically track speakers in a classroom situation, thus providing a more convenient shot of the instructor when compared to a fixed overview camera. The goal is to come to a subjective quality close to a manned camera in a standard setting, thus reducing the staffing cost for lecture recording significantly. Technically , the software consists of a set of OSGI bundles that can be added to an existing Matterhorn system or live as a stand-alone application. The video analysis portions of the system utilize the GPU through the OpenCL standard. The software breaks down into six major components: 1. a scene analysis that discovers and tracks the positions of moving objects in the scene 2. a plug-in mechanism for object analysis modules 3. a facility for communicating with PTZ cameras (currently supports Sony VISCA protocol) 4. a module that controls the movement of the PTZ camera 5. a virtual camera operator that executes camera control strategies defined in JavaScript 6. a GUI for tuning the system for individual class rooms The project is based on a bachelor thesis by Benjamin Wulff (described at http://ieeexplore.ieee.org/xpl/freeabs_all.jsp?arnumber=6123405&abstractAccess=no&userType=inst with the original project name "OpenTrack") and has Benjamin leading this effort with contributions from Alexander Fecke. LectureSight will be piloted in situ at UOS and ETH in the fall semester 2012. Expect to be informed about first results by the end of the year. For a sneak preview, take a look at the current development snapshot (http://video2.virtuos.uni-osnabrueck.de:8080/engage/ui/watch.html?id=6dba66d9-0a53-45ed-bc67-7ab3bf50b31c) or go to the project website http://lecturesight.org/. Feedback and questions are welcome. Regards RĂ¼diger, Olaf A. _______________________________________________ Community mailing list [email protected] http://lists.opencastproject.org/mailman/listinfo/community To unsubscribe please email [email protected] _______________________________________________ _______________________________________________ Matterhorn mailing list [email protected] http://lists.opencastproject.org/mailman/listinfo/matterhorn To unsubscribe please email [email protected] _______________________________________________
