计算机科学
移动边缘计算
分拆(数论)
服务器
延迟(音频)
边缘计算
能源消耗
移动设备
分布式计算
计算卸载
GSM演进的增强数据速率
方案(数学)
计算机网络
操作系统
人工智能
数学
电信
生态学
数学分析
组合数学
生物
作者
Zhuofan Liao,Weibo Hu,Jiawei Huang,Jianxin Wang
出处
期刊:Ad hoc networks
[Elsevier BV]
日期:2023-05-01
卷期号:144: 103156-103156
被引量:3
标识
DOI:10.1016/j.adhoc.2023.103156
摘要
Mobile edge computing is conducive to artificial intelligence computing near terminals, in which Deep Neural Networks (DNNs) should be partitioned to allocate tasks partially to the edge for execution to reduce latency and save energy. Most of the existing studies assume that the tasks are of the same type or the computing resources of the server are the same. In real life, Mobile Devices (MDs) and Edge Servers (ESs) are heterogeneous in type and computing resources, it is challenging to find the optimal partition point for each DNN and offload it to an appropriate ES. To fill this gap, we propose a partitioning-and-offloading scheme for the heterogeneous tasks-server system to reduce the overall system latency and energy consumption on DNN inference. The scheme has four steps. First, it establishes a partitioning and task offloading model for adaptive DNN model. Second, to reduce the solution space, the scheme designs a Partition Point Retain (PPR) algorithm. After that, the scheme gives an Optimal Partition Point (OPP) Algorithm to find the optimal partition point with the minimum cost for each ES corresponding to each MD. Based on the partition points, an offloading of DNN tasks for each MD is presented to finish the whole scheme. Simulations show that the proposed scheme reduces the total cost by 77.9% and 59.9% on average compared to Only-Local and Only-Server respectively in the heterogeneous edge computing environment.
科研通智能强力驱动
Strongly Powered by AbleSci AI