<?xml version="1.0" encoding="UTF-8"?>
<collection xmlns="http://www.loc.gov/MARC21/slim">
	<record>
		<datafield tag="980" ind1=" " ind2=" ">
			<subfield code="a">REPORT</subfield>
		</datafield>
		<datafield tag="970" ind1=" " ind2=" ">
			<subfield code="a">Jayagopi_Idiap-RR-33-2012/IDIAP</subfield>
		</datafield>
		<datafield tag="245" ind1=" " ind2=" ">
			<subfield code="a">The Vernissage Corpus: A Multimodal Human-Robot-Interaction Dataset</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Jayagopi, Dinesh Babu</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Sheikhi, Samira</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Klotz, David</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Wienke, Johannes</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Odobez, Jean-Marc</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Wrede, Sebastian</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Khalidov, Vasil</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Nguyen, Laurent Son</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Wrede, Britta</subfield>
		</datafield>
		<datafield tag="700" ind1=" " ind2=" ">
			<subfield code="a">Gatica-Perez, Daniel</subfield>
		</datafield>
		<datafield tag="088" ind1=" " ind2=" ">
			<subfield code="a">Idiap-RR-33-2012</subfield>
		</datafield>
		<datafield tag="260" ind1=" " ind2=" ">
			<subfield code="c">2012</subfield>
			<subfield code="b">Idiap</subfield>
		</datafield>
		<datafield tag="771" ind1="2" ind2=" ">
			<subfield code="d">December 2012</subfield>
		</datafield>
		<datafield tag="520" ind1=" " ind2=" ">
			<subfield code="a">We introduce a new multimodal interaction dataset with extensive annotations
in a conversational Human-Robot-Interaction (HRI) scenario. It has been recorded and 
annotated to benchmark many relevant perceptual tasks, towards 
enabling a robot to converse with multiple humans, such as   
speaker localization, key word spotting, speech recognition in audio domain;
tracking, pose estimation, nodding, visual focus of attention estimation in visual
domain; and an audio-visual task such as addressee detection. 
Some of the above mentioned tasks could benefit from information sensed from several modalities and recorded states of the robot. 
As compared to recordings done with a static camera, this corpus involves the head-movement of a humanoid 
robot (due to gaze change, nodding), 
making it challenging for tracking. Also, the significant background 
noise present in a real HRI setting makes tasks in the auditory domain more challenging. From the interaction point of view, our scenario, 
where the robot explains paintings in a room and then quizzes the participants, allows to analyze the 
quality of the interaction and the behavior of the human interaction partners.</subfield>
		</datafield>
	</record>
</collection>