Video analysis in typical applications for surveillance, multimedia content analysis and medical imaging usually concentrate on recognition of gestures, head tracking, pose/gaze and facial expression from video frames. The Video Feature Extraction CALLAS component focus instead on indicators of level of interest or enthusiasm of an audience participating to an event or installation: it uses face detection for counting and tracking people, as well as orientation of the head for head movement information and it includes also quantitative movement analysis. Processing is performed in real time by balancing the computational load between all the time consuming image processing tasks.
Availability of the component: Reference contacts are Markus Niiranen and Tommi Keränen of VTT Technical Research Centre of Finland.
Notes: The component will be made available in its compiled (32-bit Windows executable) form for non-commercial use upon request.
Besides functional tests, extensive experimentation in CALLAS is done in Scientific Showcases:
- the e-Tree : in this Augmented Reality art installation the component is used to track faces and global flow, resulting in understanding the level of attention/interest from appearance/disappearance: the event area is mapped to pleasure, and an average is performed across multiple faces, the movement in indicative of an activity (length is mapped to arousal) that recognise keywords and confidence rates.
- the Common Touch :in this collective empathic navigation of slogans on a public multitouch screen the component is used to input the number of faces in front of the installation. It might be adapted to support wide angle camera.