强化学习
学习迁移
计算机科学
人工智能
机器人
机器人学习
特征(语言学)
不变(物理)
类比
生物
人机交互
自然(考古学)
数学
移动机器人
哲学
考古
历史
语言学
数学物理
作者
Abhishek Gupta,Coline Devin,YuXuan Liu,Pieter Abbeel,Sergey Levine
出处
期刊:Cornell University - arXiv
日期:2017-01-01
被引量:112
标识
DOI:10.48550/arxiv.1703.02949
摘要
People can learn a wide range of tasks from their own experience, but can also learn from observing other creatures. This can accelerate acquisition of new skills even when the observed agent differs substantially from the learning agent in terms of morphology. In this paper, we examine how reinforcement learning algorithms can transfer knowledge between morphologically different agents (e.g., different robots). We introduce a problem formulation where two agents are tasked with learning multiple skills by sharing information. Our method uses the skills that were learned by both agents to train invariant feature spaces that can then be used to transfer other skills from one agent to another. The process of learning these invariant feature spaces can be viewed as a kind of "analogy making", or implicit learning of partial correspondences between two distinct domains. We evaluate our transfer learning algorithm in two simulated robotic manipulation skills, and illustrate that we can transfer knowledge between simulated robotic arms with different numbers of links, as well as simulated arms with different actuation mechanisms, where one robot is torque-driven while the other is tendon-driven.
科研通智能强力驱动
Strongly Powered by AbleSci AI