计算机科学
稳健性(进化)
强化学习
人工智能
循环神经网络
频道(广播)
人工神经网络
机器学习
计算机网络
生物化学
基因
化学
作者
Yue Xu,Jianyuan Yu,R. Michael Buehrer
标识
DOI:10.1109/twc.2020.2984227
摘要
This papera 1 investigates deep reinforcement learning (DRL) based on a Recurrent Neural Network (RNN) for Dynamic Spectrum Access (DSA) under partial observations, referred to as a Deep Recurrent Q-Network (DRQN). Specifically, we consider a scenario with multiple independent channels and multiple heterogeneous Primary Users (PUs). Two key challenges in our problem formulation are that we assume our DRQN node does not have any prior knowledge of the other nodes' behavior patterns and attempts to predict the future channel state based on previous observations. The goal of the DRQN is to learn a channel access strategy with a low collision rate but a high channel utilization rate. With proper definitions of the state, action and rewards, our extensive simulation results show that a DRQN-based approach can handle a variety of communication environments including dynamic environments. Further, our results show that the DRQN node is also able to cope with multi-rate and multi-agent scenarios. Importantly, we show the following benefits of using recurrent neural networks in DSA: (i) the ability to learn the optimal strategy in different environments under partial observations; (ii) robustness to imperfect observations and (iii) the ability to utilize multiple channels, and (iv) robustness in the presence of multiple agents. 1 A parton of this work was presented at MILCOM 2018 in [1].
科研通智能强力驱动
Strongly Powered by AbleSci AI