强化学习
计算机科学
增强学习
马尔可夫链
干扰(通信)
分布式计算
频道(广播)
马尔可夫过程
计算机网络
光谱(功能分析)
传输(电信)
频谱管理
闲置
无线网络
无线
人工智能
认知无线电
电信
机器学习
数学
操作系统
物理
统计
量子力学
作者
Manish Anand Yadav,Yuhui Li,Guangjin Fang,Bin Shen
标识
DOI:10.1109/ccai55564.2022.9807797
摘要
To solve the problem of spectrum scarcity and spectrum under-utilization in wireless networks, we propose a double deep Q-network based reinforcement learning algorithm for distributed dynamic spectrum access. Channels in the network are either busy or idle based on the two-state Markov chain. At the start of each time slot, every secondary user (SU) performs spectrum sensing on each channel and accesses one based on the sensing result as well as the output of the Q-network of our algorithm. Over time, the Deep Reinforcement Learning (DRL) algorithm learns the spectrum environment and becomes good at modeling the behavior pattern of the primary users (PUs). Through simulation, we show that our proposed algorithm is simple to train, yet effective in reducing interference to primary as well as secondary users and achieving higher successful transmission.
科研通智能强力驱动
Strongly Powered by AbleSci AI