DDPQN: An Efficient DNN Offloading Strategy in Local-Edge-Cloud Collaborative Environments

Published: 01 Jan 2022, Last Modified: 28 Sept 2024IEEE Trans. Serv. Comput. 2022EveryoneRevisionsBibTeXCC BY-SA 4.0
Abstract: With the rapid development of the Internet of Things (IoT) and communication technology, Deep Neural Network (DNN) applications like computer vision, can now be widely used in IoT devices. However, due to the insufficient memory, low computing capacity, and low battery capacity of IoT devices, it is difficult to support the high-efficiency DNN inference and meet users’ requirements for Quality of Service (QoS). Worse still, offloading failures may occur during the massive DNN data transmission due to the intermittent wireless connectivity between IoT devices and the cloud. In order to fill this gap, we consider the partitioning and offloading of the DNN model, and design a novel optimization method for parallel offloading of large-scale DNN models in a local-edge-cloud collaborative environment with limited resources. Combined with the coupling coordination degree and node balance degree, an improved Double Dueling Prioritized deep Q-Network (DDPQN) algorithm is proposed to obtain the DNN offloading strategy. Compared with existing algorithms, the DDPQN algorithm can obtain an efficient DNN offloading strategy with low delay, low energy consumption, and low cost under the premise of ensuring “delay-energy-cost” coordination and reasonable allocation of computing resources in a local-edge-cloud collaborative environment.
Loading