CONF
lathoud04c/IDIAP
AV16.3: an Audio-Visual Corpus for Speaker Localization and Tracking
Lathoud, Guillaume
Odobez, Jean-Marc
Gatica-Perez, Daniel
EXTERNAL
https://publications.idiap.ch/attachments/papers/2004/lathoud04c.pdf
PUBLIC
https://publications.idiap.ch/index.php/publications/showcite/lathoud-rr-04-28
Related documents
Proceedings of the 2004 MLMI Workshop, S. Bengio and H. Bourlard Eds, Springer Verlag
2005
IDIAP-RR 04-28
Assessing the quality of a speaker localization or tracking algorithm on a few short examples is difficult, especially when the ground-truth is absent or not well defined. One step towards systematic performance evaluation of such algorithms is to provide time-continuous speaker location annotation over a series of real recordings, covering various test cases. Areas of interest include audio, video and audio-visual speaker localization and tracking. The desired location annotation can be either 2-dimensional (image plane) or 3-dimensional (physical space). This paper motivates and describes a corpus of audio-visual data called ``AV16.3'', along with a method for 3-D location annotation based on calibrated cameras. ``16.3'' stands for 16 microphones and 3 cameras, recorded in a fully synchronized manner, in a meeting room. Part of this corpus has already been successfully used to report research results.
REPORT
lathoud-rr-04-28/IDIAP
AV16.3: an Audio-Visual Corpus for Speaker Localization and Tracking
Lathoud, Guillaume
Odobez, Jean-Marc
Gatica-Perez, Daniel
EXTERNAL
https://publications.idiap.ch/attachments/reports/2004/rr-04-28.pdf
PUBLIC
Idiap-RR-28-2004
2004
IDIAP
Martigny, Switzerland
Published in ``Proceedings of the 2004 MLMI Workshop''
Assessing the quality of a speaker localization or tracking algorithm on a few short examples is difficult, especially when the ground-truth is absent or not well defined. One step towards systematic performance evaluation of such algorithms is to provide time-continuous speaker location annotation over a series of real recordings, covering various test cases. Areas of interest include audio, video and audio-visual speaker localization and tracking. The desired location annotation can be either 2-dimensional (image plane) or 3-dimensional (physical space). This paper motivates and describes a corpus of audio-visual data called ``AV16.3'', along with a method for 3-D location annotation based on calibrated cameras. ``16.3'' stands for 16 microphones and 3 cameras, recorded in a fully synchronized manner, in a meeting room. Part of this corpus has already been successfully used to report research results.