强化学习
运动规划
机器人
稳健性(进化)
计算机科学
碰撞
抓住
循环神经网络
人工神经网络
人工智能
路径(计算)
模拟
地形
路径长度
数学优化
数学
地理
基因
地图学
生物化学
计算机安全
化学
程序设计语言
计算机网络
作者
Guichao Lin,Lixue Zhu,Jinhui Li,Xiangjun Zou,Yunchao Tang
标识
DOI:10.1016/j.compag.2021.106350
摘要
In unstructured orchard environments, picking a target fruit without colliding with neighboring branches is a significant challenge for guava-harvesting robots. This paper introduces a fast and robust collision-free path-planning method based on deep reinforcement learning. A recurrent neural network is first adopted to remember and exploit the past states observed by the robot, then a deep deterministic policy gradient algorithm (DDPG) predicts a collision-free path from the states. A simulation environment is developed and its parameters are randomized during the training phase to enable recurrent DDPG to generalize to real-world scenarios. We also introduce an image processing method that uses a deep neural network to detect obstacles and uses many three-dimensional line segments to approximate the obstacles. Simulations show that recurrent DDPG only needs 29 ms to plan a collision-free path with a success rate of 90.90%. Field tests show that recurrent DDPG can increase grasp, detachment, and harvest success rates by 19.43%, 9.11%, and 10.97%, respectively, compared to cases where no collision-free path-planning algorithm is implemented. Recurrent DDPG strikes a strong balance between efficiency and robustness and may be suitable for other fruits.
科研通智能强力驱动
Strongly Powered by AbleSci AI