强化学习
计算机科学
动作(物理)
水下
建筑
状态空间
海洋哺乳动物
噪音(视频)
功能(生物学)
形势意识
实时计算
空格(标点符号)
信号(编程语言)
人工智能
工程类
数学
地理
统计
物理
考古
量子力学
进化生物学
渔业
航空航天工程
图像(数学)
生物
程序设计语言
操作系统
作者
Edward Clark,Alan J. Hunter,Olga Isupova,Marcus Donnelly
出处
期刊:Proceedings of Meetings on Acoustics
日期:2022-01-01
卷期号:47: 070008-070008
摘要
Underwater passive acoustic source detection and tracking is important for various marine applications, including marine mammal monitoring and naval surveillance. The performance in these applications is dependent on the placement and operation of sensing assets, such as autonomous underwater vehicles. Conventionally, these decisions have been made by human operators aided by acoustic propagation modelling tools, situational and environmental data, and experience. However, this is time-consuming and computationally expensive. We consider a 'toy problem' of a single autonomous vehicle (agent) in search of a stationary source of low frequency within a reinforcement learning (RL) architecture. We initially choose the observation space to be the agent's current position. The agent is allowed to explore the environment with a limited action space, taking equal distance steps in one of $n$ directions. Rewards are received for positive detections of the source. Using OpenAI's PPO algorithm an increase in median episode reward of approximately 20 points in the RL environment developed is seen when the agent is given a history of it's previous moves and signal-to-noise ratio compared to the simple state. The future expansion of the RL framework is discussed in terms of the observation and action spaces, reward function and RL architecture.
科研通智能强力驱动
Strongly Powered by AbleSci AI