@inproceedings{262c2208df754948b8b24ba7f763e235,
title = "SRG3: Speech-driven Robot Gesture Generation with GAN",
abstract = "The human gestures occur spontaneously and usually they are aligned with speech, which leads to a natural and expressive interaction. Speech-driven gesture generation is important in order to enable a social robot to exhibit social cues and conduct a successful human-robot interaction. In this paper, the generation process involves mapping acoustic speech representation to the corresponding gestures for a humanoid robot. The paper proposes a new GAN (Generative Adversarial Network) architecture for speech to gesture generation. Instead of the fixed mapping from one speech to one gesture pattern, our end-to-end GAN structure can generate multiple mapped gestures patterns from one speech (with multiple noises) just like humans do. The generated gestures can be applied to social robots with arms. The evaluation result shows the effectiveness of our generative model for speech-driven robot gesture generation.",
author = "Chuang Yu and Adriana Tapus",
note = "Publisher Copyright: {\textcopyright} 2020 IEEE.; 16th IEEE International Conference on Control, Automation, Robotics and Vision, ICARCV 2020 ; Conference date: 13-12-2020 Through 15-12-2020",
year = "2020",
month = dec,
day = "13",
doi = "10.1109/ICARCV50220.2020.9305330",
language = "English",
series = "16th IEEE International Conference on Control, Automation, Robotics and Vision, ICARCV 2020",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "759--766",
booktitle = "16th IEEE International Conference on Control, Automation, Robotics and Vision, ICARCV 2020",
}