已入深夜,您辛苦了!由于当前在线用户较少,发布求助请尽量完整地填写文献信息,科研通机器人24小时在线,伴您度过漫漫科研夜!祝你早点完成任务,早点休息,好梦!

Energy-Efficient Trajectory Optimization with Wireless Charging in UAV-Assisted MEC Based on Multi-Objective Reinforcement Learning

强化学习 计算机科学 弹道 无线 轨迹优化 能量(信号处理) 实时计算 人工智能 电信 天文 数学 统计 物理
作者
Fuhong Song,Mingsen Deng,Huanlai Xing,Yanping Liu,Fei Ye,Zhiwen Xiao
出处
期刊:IEEE Transactions on Mobile Computing [IEEE Computer Society]
卷期号:: 1-18 被引量:7
标识
DOI:10.1109/tmc.2024.3384405
摘要

This paper investigates the problem of energy-efficient trajectory optimization with wireless charging (ETWC) in an unmanned aerial vehicle (UAV)-assisted mobile edge computing system. A UAV is dispatched to collect computation tasks from specific ground smart devices (GSDs) within its coverage while transmitting energy to the other GSDs. In addition, a high-altitude platform with a laser beam is deployed in the stratosphere to charge the UAV, so as to maintain its flight mission. The ETWC problem is characterized by multi-objective optimization, aiming to maximize both the energy efficiency of the UAV and the number of tasks collected via optimizing the UAV's flight trajectories. The conflict between the two objectives in the problem makes it quite challenging. Recently, some single-objective reinforcement learning (SORL) algorithms have been introduced to address the aforementioned problem. Nevertheless, these SORLs adopt linear scalarization to define the user utility, thus ignoring the conflict between objectives. Furthermore, in dynamic MEC scenarios, the relative importance assigned to each objective may vary over time, posing significant challenges for conventional SORLs. To solve the challenge, we first build a multi-objective Markov decision process that has a vectorial reward mechanism. There is a corresponding relationship between each component of the reward and one of the two objectives. Then, we propose a new trace-based experience replay scheme to modify sample efficiency and reduce replay buffer bias, resulting in a modified multi-objective reinforcement learning algorithm. The experiment results validate that the proposed algorithm can obtain better adaptability to dynamic preferences and a more favorable balance between objectives compared with several algorithms.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
NexusExplorer应助Zr采纳,获得20
1秒前
小二郎应助666采纳,获得10
1秒前
隐形曼青应助小赵采纳,获得10
1秒前
3秒前
方若剑发布了新的文献求助10
3秒前
无极微光应助蓝天下载采纳,获得20
4秒前
5秒前
fanjinze完成签到,获得积分10
5秒前
mangfu发布了新的文献求助10
5秒前
6秒前
7秒前
paul完成签到,获得积分10
8秒前
刘丰铭发布了新的文献求助10
9秒前
9秒前
嘿嘿完成签到 ,获得积分10
10秒前
敏感思山完成签到 ,获得积分10
10秒前
best发布了新的文献求助10
12秒前
lq发布了新的文献求助10
13秒前
666发布了新的文献求助10
14秒前
bailubailing发布了新的文献求助10
15秒前
uu完成签到 ,获得积分10
16秒前
生动的水云完成签到,获得积分10
21秒前
科研通AI6.2应助wang采纳,获得30
21秒前
李兴完成签到 ,获得积分10
22秒前
shen完成签到 ,获得积分10
23秒前
一川完成签到,获得积分10
23秒前
FashionBoy应助发呆小绵羊采纳,获得10
23秒前
Zzc2026应助Zr采纳,获得40
24秒前
bai完成签到 ,获得积分10
25秒前
26秒前
慕青应助科研通管家采纳,获得10
26秒前
Nexus应助科研通管家采纳,获得10
26秒前
Nexus应助科研通管家采纳,获得10
26秒前
26秒前
香蕉觅云应助科研通管家采纳,获得10
26秒前
在水一方应助科研通管家采纳,获得10
27秒前
丘比特应助科研通管家采纳,获得10
27秒前
27秒前
27秒前
CipherSage应助科研通管家采纳,获得10
27秒前
高分求助中
Standards for Molecular Testing for Red Cell, Platelet, and Neutrophil Antigens, 7th edition 1000
HANDBOOK OF CHEMISTRY AND PHYSICS 106th edition 1000
ASPEN Adult Nutrition Support Core Curriculum, Fourth Edition 1000
Signals, Systems, and Signal Processing 610
脑电大模型与情感脑机接口研究--郑伟龙 500
GMP in Practice: Regulatory Expectations for the Pharmaceutical Industry 500
简明药物化学习题答案 500
热门求助领域 (近24小时)
化学 材料科学 医学 生物 纳米技术 工程类 有机化学 化学工程 生物化学 计算机科学 物理 内科学 复合材料 催化作用 物理化学 光电子学 电极 细胞生物学 基因 无机化学
热门帖子
关注 科研通微信公众号,转发送积分 6298841
求助须知:如何正确求助?哪些是违规求助? 8115759
关于积分的说明 16990365
捐赠科研通 5360089
什么是DOI,文献DOI怎么找? 2847564
邀请新用户注册赠送积分活动 1825013
关于科研通互助平台的介绍 1679320