强化学习
仿人机器人
敏捷软件开发
投掷
钢筋
人工智能
球(数学)
机器人
模拟
计算机科学
工程类
人机交互
结构工程
航空学
软件工程
数学分析
数学
作者
Tuomas Haarnoja,Ben Moran,Guy Lever,Sandy H. Huang,Dhruva Tirumala,Jan Humplik,Markus Wulfmeier,Saran Tunyasuvunakool,Noah Siegel,Roland Hafner,Michael Bloesch,Kristian Hartikainen,Arunkumar Byravan,Leonard Hasenclever,Yuval Tassa,Fereshteh Sadeghi,Nathan Batchelor,Federico Casarini,Stefano Saliceti,Charles Game
出处
期刊:Science robotics
[American Association for the Advancement of Science]
日期:2024-04-10
卷期号:9 (89)
被引量:33
标识
DOI:10.1126/scirobotics.adi8022
摘要
We investigated whether deep reinforcement learning (deep RL) is able to synthesize sophisticated and safe movement skills for a low-cost, miniature humanoid robot that can be composed into complex behavioral strategies. We used deep RL to train a humanoid robot to play a simplified one-versus-one soccer game. The resulting agent exhibits robust and dynamic movement skills, such as rapid fall recovery, walking, turning, and kicking, and it transitions between them in a smooth and efficient manner. It also learned to anticipate ball movements and block opponent shots. The agent’s tactical behavior adapts to specific game contexts in a way that would be impractical to manually design. Our agent was trained in simulation and transferred to real robots zero-shot. A combination of sufficiently high-frequency control, targeted dynamics randomization, and perturbations during training enabled good-quality transfer. In experiments, the agent walked 181% faster, turned 302% faster, took 63% less time to get up, and kicked a ball 34% faster than a scripted baseline.
科研通智能强力驱动
Strongly Powered by AbleSci AI