Multimodal Integration for Meeting Group Action Segmentation and Recognition
| Type of publication: | Idiap-RR |
| Citation: | zhang-rr-05-31 |
| Number: | Idiap-RR-31-2005 |
| Year: | 2005 |
| Institution: | IDIAP |
| Address: | Martigny, Switzerland |
| Note: | Published in ``MLMI'', July, 2005 |
| Abstract: | We address the problem of segmentation and recognition of sequences of multimodal human interactions in meetings. These interactions can be seen as a rough structure of a meeting, and can be used either as input for a meeting browser or as a first step towards a higher semantic analysis of the meeting. A common lexicon of multimodal group meeting actions, a shared meeting data set, and a common evaluation procedure enable us to compare the different approaches. We compare three different multimodal feature sets and four modelling infrastructures: a higher semantic feature approach, multi-layer HMMs, a multi-stream DBN, as well as a multi-stream mixed-state DBN for disturbed data. |
| Userfields: | ipdinar={2005}, ipdmembership={vision}, language={English}, |
| Keywords: | |
| Projects: |
Idiap |
| Authors: | |
| Crossref by |
zhang-rr-05-31b |
| Added by: | [UNK] |
| Total mark: | 0 |
|
Attachments
|
|
|
Notes
|
|
|
|
|