RAVEL: an annotated corpus for training robots with audiovisual abilities. Alameda-Pineda, X., Sanchez-Riera, J., Wienke, J., Franc, V., Čech, J., Kulkarni, K., Deleforge, A., & Horaud, R. Journal on Multimodal User Interfaces, 7(1-2):79--91, March, 2013.
RAVEL: an annotated corpus for training robots with audiovisual abilities [link]Paper  doi  abstract   bibtex   
We introduce Ravel (Robots with Audiovisual Abilities), a publicly available data set which covers examples of Human Robot Interaction (HRI) scenarios. These scenarios are recorded using the audio-visual robot head POPEYE, equipped with two cameras and four microphones, two of which being plugged into the ears of a dummy head. All the recordings were performed in a standard room with no special equipment, thus providing a challenging indoor scenario. This data set provides a basis to test and benchmark methods and algorithms for audio-visual scene analysis with the ultimate goal of enabling robots to interact with people in the most natural way. The data acquisition setup, sensor calibration, data annotation and data content are fully detailed. Moreover, three examples of using the recorded data are provided, illustrating its appropriateness for carrying out a large variety of HRI experiments. The Ravel data are publicly available at: http://ravel.humavips.eu/.
@article{alameda-pineda_ravel:_2013,
	title = {{RAVEL}: an annotated corpus for training robots with audiovisual abilities},
	volume = {7},
	issn = {1783-7677, 1783-8738},
	shorttitle = {{RAVEL}},
	url = {http://link.springer.com/article/10.1007/s12193-012-0111-y},
	doi = {10.1007/s12193-012-0111-y},
	abstract = {We introduce Ravel (Robots with Audiovisual Abilities), a publicly available data set which covers examples of Human Robot Interaction (HRI) scenarios. These scenarios are recorded using the audio-visual robot head POPEYE, equipped with two cameras and four microphones, two of which being plugged into the ears of a dummy head. All the recordings were performed in a standard room with no special equipment, thus providing a challenging indoor scenario. This data set provides a basis to test and benchmark methods and algorithms for audio-visual scene analysis with the ultimate goal of enabling robots to interact with people in the most natural way. The data acquisition setup, sensor calibration, data annotation and data content are fully detailed. Moreover, three examples of using the recorded data are provided, illustrating its appropriateness for carrying out a large variety of HRI experiments. The Ravel data are publicly available at: http://ravel.humavips.eu/.},
	language = {en},
	number = {1-2},
	urldate = {2014-05-19TZ},
	journal = {Journal on Multimodal User Interfaces},
	author = {Alameda-Pineda, Xavier and Sanchez-Riera, Jordi and Wienke, Johannes and Franc, Vojtěch and Čech, Jan and Kulkarni, Kaustubh and Deleforge, Antoine and Horaud, Radu},
	month = mar,
	year = {2013},
	pages = {79--91}
}

Downloads: 0