Towards a multimodal meeting record - Robotics Institute Carnegie Mellon University

Towards a multimodal meeting record

Ralph Gross, Michael Bett, H. Yu, Xu Zhu, Y. Pan, Jie Yang, and Alex Waibel
Conference Paper, Proceedings of IEEE International Conference on Multimedia and Expo (ICME '00), Vol. 3, pp. 1593 - 1596, July, 2000

Abstract

Face-to-face meetings usually encompass several modalities including speech, gesture, handwriting, and person identification. Recognition and integration of each of these modalities is important to create an accurate record of a meeting. However, each of these modalities presents recognition difficulties. Speech recognition must be speaker and domain independent, have low word error rates, and be close to real time to be useful. Gesture and handwriting recognition must be writer independent and support a wide variety of writing styles. Person identification has difficulty with segmentation in a crowded room. Furthermore, in order to produce the record automatically, we have to solve the assignment problem (who is saying what), which involves people identification and speech recognition. This paper examines a multimodal meeting room system under development at Carnegie Mellon University that enables us to track, capture and integrate the important aspects of a meeting from people identification to meeting transcription. Once a multimedia meeting record is created, it can be archived for later retrieval.

BibTeX

@conference{Gross-2000-8090,
author = {Ralph Gross and Michael Bett and H. Yu and Xu Zhu and Y. Pan and Jie Yang and Alex Waibel},
title = {Towards a multimodal meeting record},
booktitle = {Proceedings of IEEE International Conference on Multimedia and Expo (ICME '00)},
year = {2000},
month = {July},
volume = {3},
pages = {1593 - 1596},
}