Towards More Realistic Human-Robot Conversation: A Seq2seq-Based Body Gesture Interaction System

2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS)(2019)

Cited 11|Views2
No score
Abstract
This paper presents a novel system that enables intelligent robots to exhibit realistic body gestures while communicating with humans. The proposed system consists of a listening model and a speaking model used in corresponding conversational phases. Both models are adapted from the sequence-to-sequence (seq2seq) architecture to synthesize body gestures represented by the movements of twelve upper-body keypoints. All the extracted 2D keypoints are firstly 3D-transformed, then rotated and normalized to discard irrelevant information. Substantial videos of human conversations from Youtube are collected and preprocessed to train the listening and speaking models separately, after which the two models are evaluated using metrics of mean squared error (MSE) and cosine similarity on the test dataset. The tuned system is implemented to drive a virtual avatar as well as Pepper, a physical humanoid robot, to demonstrate the improvement on conversational interaction abilities of our method in practice.
More
Translated text
Key words
physical humanoid robot,conversational interaction abilities,realistic human-robot conversation,seq2seq-based body gesture interaction system,intelligent robots,realistic body gestures,listening model,speaking model,sequence-to-sequence architecture,synthesize body gestures,upper-body keypoints,extracted 2D keypoints,human conversations,listening speaking models,tuned system,conversational phases
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined