Vision-Based Multi-Task Manipulation for Inexpensive Robots Using End-To-End Learning from Demonstration

2018 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA)(2018)

引用 254|浏览133
暂无评分
摘要
We propose a technique for multi-task learning from demonstration that trains the controller of a low-cost robotic arm to accomplish several complex picking and placing tasks, as well as non-prehensile manipulation. The controller is a recurrent neural network using raw images as input and generating robot arm trajectories, with the parameters shared across the tasks. The controller also combines VAE-GAN-based reconstruction with autoregressive multimodal action prediction. Our results demonstrate that it is possible to learn complex manipulation tasks, such as picking up a towel, wiping an object, and depositing the towel to its previous position, entirely from raw images with direct behavior cloning. We show that weight sharing and reconstruction-based regularization substantially improve generalization and robustness, and training on multiple tasks simultaneously increases the success rate on all tasks.
更多
查看译文
关键词
nonprehensile manipulation,recurrent neural network,raw images,VAE-GAN-based reconstruction,autoregressive multimodal action prediction,complex manipulation tasks,towel,weight,reconstruction-based regularization,vision-based multitask manipulation,end-to-end learning,multitask learning,low-cost robotic arm,robot arm trajectories,complex picking and placing tasks
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要