Interpreting Multimodal Communication Scenes
Add to your list(s)
Download to your calendar using vCal
If you have a question about this talk, please contact Dr Marcus Tomalin.
The AMI project is about interpreting human communication using
audio, video and other signals, in the context of multiparty
meetings. This is an interdisciplinary problem involving signal
processing and machine learning (how to make sense of multimodal
communication scenes), understanding the content of the recognised
signals, and developing useful applications.
In this talk I shall discuss some of the work we have been doing to
automatically annotated and interpret multimodal meeting recordings.
Specific issues that I’ll discuss include the use of multistream
statistical models to segment meetings at different levels (dialogue
acts and ``meeting actions’‘) and approaches to meeting summarisation.
This talk is part of the Machine Intelligence Laboratory Speech Seminars series.
This talk is included in these lists:
Note that ex-directory lists are not shown.
|