Audio-Visual Processing in Meetings: Seven Questions and Current AMI Answers

Autor: Jan Cernocky, Andrew Thean, Ronald Müller, Hervé Bourlard, Fabien Cardinaux, Pavel Zemcik, Marc Al-Hames, Mannes Poel, Jean-Marc Odobez, Silèye O. Ba, Daniel Gatica-Perez, Steve Renals, David A. van Leeuwen, Petr Motlicek, Kevin Smith, Jeroen van Rest, Gerhard Rigoll, Thomas Hain, Sébastien Marcel, Sascha Schreiber, Stephan Reiter, Adam Janin, Rutger Rienks
Rok vydání: 2007
Předmět:
Zdroj: Machine Learning for Multimodal Interaction ISBN: 9783540692676
MLMI
Proceedings of the 3rd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006), 24-35
STARTPAGE=24;ENDPAGE=35;TITLE=Proceedings of the 3rd Joint Workshop on Multimodal Interaction and Related Machine Learning Algorithms (MLMI 2006)
ISSN: 0302-9743
DOI: 10.1007/11965152_3
Popis: The project Augmented Multi-party Interaction (AMI) is concerned with the development of meeting browsers and remote meeting assistants for instrumented meeting rooms – and the required component technologies R&D themes: group dynamics, audio, visual, and multimodal processing, content abstraction, and human-computer interaction. The audio-visual processing workpackage within AMI addresses the automatic recognition from audio, video, and combined audio-video streams, that have been recorded during meetings. In this article we describe the progress that has been made in the first two years of the project. We show how the large problem of audio-visual processing in meetings can be split into seven questions, like “Who is acting during the meeting?��?. We then show which algorithms and methods have been developed and evaluated for the automatic answering of these questions.
Databáze: OpenAIRE