Learning Non-verbal Behavior for a Social Robot from YouTube Videos. Jonell, P., Kucherenko, T., Ekstedt, E., & Beskow, J. In ICDL-EpiRob Workshop on Naturalistic Non-Verbal and Affective Human-Robot Interactions, 2019.
Learning Non-verbal Behavior for a Social Robot from YouTube Videos [link]Website  abstract   bibtex   
Non-verbal behavior is crucial for positive perception of humanoid robots. If modeled well it can improve the interaction and leave the user with a positive experience, on the other hand, if it is modelled poorly it may impede the interaction and become a source of distraction. Most of the existing work on modeling non-verbal behavior show limited variability due to the fact that the models employed are deterministic and the generated motion can be perceived as repetitive and predictable. In this paper, we present a novel method for generation of a limited set of facial expressions and head movements, based on a probabilistic generative deep learning architecture called Glow. We have implemented a workflow which takes videos directly from YouTube, extracts relevant features, and trains a model that generates gestures that can be realized in a robot without any post processing. A user study was conducted and illustrated the importance of having any kind of non-verbal behavior while most differences between the ground truth, the proposed method, and a random control were not significant (however, the differences that were significant were in favor of the proposed method).
@inproceedings{
 title = {Learning Non-verbal Behavior for a Social Robot from YouTube Videos},
 type = {inproceedings},
 year = {2019},
 keywords = {Facial expressions,generative models,head movement,neural network,non-verbal behavior,social robotics},
 websites = {https://nicolas-navarro-guerrero.gitlab.io/workshop-non-verbal-human-robot-interactions-icdl-epirob-2019/},
 institution = {KTH, Speech, Music and Hearing},
 id = {8640b02c-ffac-3d0d-b1aa-6689a1d57352},
 created = {2020-01-06T19:15:01.816Z},
 file_attached = {false},
 profile_id = {fb8d345a-1d79-3791-a6c6-00233ea44521},
 last_modified = {2020-07-06T09:08:48.686Z},
 read = {false},
 starred = {false},
 authored = {true},
 confirmed = {true},
 hidden = {false},
 citation_key = {Jonell1357431},
 source_type = {inproceedings},
 notes = {QC 20191007},
 folder_uuids = {b9e8b2b5-59ea-4ffd-818c-22a0d98c2144,f59b6e50-2bf8-4a1f-bfbb-f7865b907fde},
 private_publication = {false},
 abstract = {Non-verbal behavior is crucial for positive perception of humanoid robots. If modeled well it can improve the interaction and leave the user with a positive experience, on the other hand, if it is modelled poorly it may impede the interaction and become a source of distraction. Most of the existing work on modeling non-verbal behavior show limited variability due to the fact that the models employed are deterministic and the generated motion can be perceived as repetitive and predictable. In this paper, we present a novel method for generation of a limited set of facial expressions and head movements, based on a probabilistic generative deep learning architecture called Glow. We have implemented a workflow which takes videos directly from YouTube, extracts relevant features, and trains a model that generates gestures that can be realized in a robot without any post processing. A user study was conducted and illustrated the importance of having any kind of non-verbal behavior while most differences between the ground truth, the proposed method, and a random control were not significant (however, the differences that were significant were in favor of the proposed method).},
 bibtype = {inproceedings},
 author = {Jonell, Patrik and Kucherenko, Taras and Ekstedt, Erik and Beskow, Jonas},
 booktitle = {ICDL-EpiRob Workshop on Naturalistic Non-Verbal and Affective Human-Robot Interactions}
}

Downloads: 0