Intelligently Integrating Information from Speech and Vision Processing to Perform Light-weight Meeting Understanding

Alexander Rudnicky, Paul Rybski, Satanjeev Banerjee, and Manuela Veloso
International Conference on Multi-modal Interfaces (ICMI), International Workshop on Multimodal Multiparty Meeting Processing , October, 2005.


Download
  • Adobe portable document format (pdf) (558KB)
Copyright notice: This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. These works may not be reposted without the explicit permission of the copyright holder.

Abstract
Important information is often generated at meetings but identifying, and retrieving that information after the meeting is not always simple. Automatically capturing such information and making it available for later retrieval has therefore become a topic of some interest. Most approaches to this problem have involved constructing specialized instrumented meeting rooms that allow a meeting to be captured in great detail. We propose an alternate approach that focuses on people’s information retrieval needs and makes use of a light-weight data collection system that allows data acquisition on portable equipment, such as personal laptops. Issues that arise include the integration of information from different audio and video streams and optimumuse of sparse computing resources. This paper describes our current development of a light-weight portable meeting recording infrastructure, as well as the use of streams of visual and audio information to derive structure from meetings. The goal is to make meeting contents easily accessible to people.

Notes
Sponsor: National Business Center (NBC) of the Department of the Interior (DOI) under a subcontract from SRI International.
Associated Center(s) / Consortia: Vision and Autonomous Systems Center
Associated Lab(s) / Group(s): People Image Analysis Consortium and MultiRobot Lab
Associated Project(s): Camera Assisted Meeting Event Observer
Number of pages: 6

Text Reference
Alexander Rudnicky, Paul Rybski, Satanjeev Banerjee, and Manuela Veloso, "Intelligently Integrating Information from Speech and Vision Processing to Perform Light-weight Meeting Understanding," International Conference on Multi-modal Interfaces (ICMI), International Workshop on Multimodal Multiparty Meeting Processing , October, 2005.

BibTeX Reference
@inproceedings{Rudnicky_2005_6659,
   author = "Alexander Rudnicky and Paul Rybski and Satanjeev Banerjee and Manuela Veloso",
   title = "Intelligently Integrating Information from Speech and Vision Processing to Perform Light-weight Meeting Understanding",
   booktitle = "International Conference on Multi-modal Interfaces (ICMI), International Workshop on Multimodal Multiparty Meeting Processing ",
   month = "October",
   year = "2005",
}