Author Login Editor-in-Chief Peer Review Editor Work Office Work

Computer Engineering ›› 2022, Vol. 48 ›› Issue (11): 30-38. doi: 10.19678/j.issn.1000-3428.0063579

• Research Hotspots and Reviews • Previous Articles     Next Articles

Cloud-Edge Collaborative DNN Inference Based on Deep Reinforcement Learning

LIU Xianfeng, LIANG Sai, LI Qiang, ZHANG Jin   

  1. College of Information Science and Engineering, Hunan Normal University, Changsha 410081, China
  • Received:2022-03-31 Revised:2022-06-05 Published:2022-06-30

基于深度强化学习的云边协同DNN推理

刘先锋, 梁赛, 李强, 张锦   

  1. 湖南师范大学 信息科学与工程学院, 长沙 410081
  • 作者简介:刘先锋(1964—),男,教授、博士,主研方向为边缘计算、云计算、人工智能;梁赛,硕士研究生;李强(通信作者),副教授、博士;张锦,教授、博士。
  • 基金资助:
    湖南省自然科学基金(2021JJ30456);湖南省科技计划(2021GK5014,2019SK2161,2018TP1018)。

Abstract: In the existing Deep Neural Network(DNN) inference based on cloud-edge collaboration, only the static partition strategy is considered in the homogeneous case of edge devices.However, the influence of the network transmission rate, edge device resources, cloud server load on the optimal partition point of DNN inference computation and the optimal offloading strategy of DNN inference task are not considered in heterogeneous edge device clusters.To solve these problems, this study presents an adaptive DNN inference computation partition and task offloading algorithm based on Deep Reinforcement Learning(DRL).The aim is to minimize the DNN inference delay, and a mathematical model of adaptive task offloading and DNN inference computation partition is established.The state, action space, and reward are defined to transform task offloading and DNN inference computation partition combination optimization problems into the optimal policy problem under the Markov decision process.DRL is used to learn from the experience pool about the approximate optimal strategy of DNN inference computation partition between edge devices and cloud servers and task offloading between heterogeneous edge clusters in a dynamic environment.The experimental results show that compared with several classical DNN inference algorithms, the DNN inference delay of the proposed algorithm in a heterogeneous dynamic environment is reduced by approximately 28.83% on average, proving that the low latency requirement of DNN inference is met in a better manner.

Key words: edge computing, Deep Neural Network(DNN), Deep Reinforcement Learning(DRL), inference computation partition, task offloading

摘要: 现有基于云边协同的深度神经网络(DNN)推理仅涉及边缘设备同构情况下的静态划分策略,未考虑网络传输速率、边缘设备资源、云服务器负载等变化对DNN推理计算最佳划分点的影响,以及异构边缘设备集群间DNN推理任务的最佳卸载策略。针对以上问题,提出基于深度强化学习的自适应DNN推理计算划分和任务卸载算法。以最小化DNN推理时延为优化目标,建立自适应DNN推理计算划分和任务卸载的数学模型。通过定义状态、动作空间和奖励,将DNN推理计算划分和任务卸载组合优化问题转换为马尔可夫决策过程下的最优策略问题。利用深度强化学习方法,从经验池中学习动态环境下边缘设备与云服务器间DNN推理计算划分和异构边缘集群间任务卸载的近似最优策略。实验结果表明,与经典DNN推理算法相比,该算法在异构动态环境下的DNN推理时延约平均降低了28.83%,能更好地满足DNN推理的低时延需求。

关键词: 边缘计算, 深度神经网络, 深度强化学习, 推理计算划分, 任务卸载

CLC Number: