IM2 (www.im2.ch) is one the 20 Swiss National Centres of Competence in Research (NCCR) aiming at boosting research and development in several areas considered of strategic importance to the Swiss economy. The National Centers of Competence in Research are a research instrument managed by the Swiss National Science Foundation (SNSF) on behalf of the Federal Authorities. Granted for a maximum duration of 12 years, they are evaluated every year by a review panel, and renewed every four years. In December 2009, the SNSF approved the next and last four-year period (2010 - 2013) which has started January 1st, 2010. Success of the NCCRs is measured in terms of research achievements, training of young scientists (PhD students and postdocs), knowledge and technology transfer (including spin-offs), and advancement of women.
The IM2 NCCR, headed by Idiap Research Institute in Martigny (www.idiap.ch), gathers the major Swiss contributors to
the field of Interactive Multimodal Information Management, from a number of
university institutions (Ecole Polytechnique Fédérale de Lausanne,
Started in 2002, IM2 is aimed at the advancement of research, and the development of prototypes, in the field of multimodal information management and man-machine interaction. IM2 is particularly concerned with technologies coordinating natural input modes (such as speech, image, pen, touch, hand gestures, head and/or body movements, and even physiological signals) with multimedia system outputs, such as speech, sounds, images, animations, whether standalone or combined in web-based browsing interfaces.
To foster collaboration and maximize synergy between the IM2 partners, it was decided to focus on a few, well defined, applications. The most important and challenging of them is Smart Meeting Management. The overall objective of this application is the construction of a demonstration system to enable structuring, browsing and querying of an archive of automatically analysed meetings. The archived meetings take place in a room equipped with multimodal sensors. For each meeting, audio, video and textual information is available, as well as information from other modalities. Audio information comes from close talking and distant microphones, as well as binaural recordings. Video information comes from multiple cameras and overhead projectors. Writing devices capture whiteboard or participant’s notes. The video and audio information form several streams of data generated during the meeting, while the additional information — the agenda, discussion papers — and can be used to guide the automatic structuring of the meeting. Multimodal interfaces called ‘meeting browsers’ are then used for searching and browsing the multimedia/multimodal meeting archive.
Last modified 2010-01-20 09:43