Abstract: With the boom in fifth-generation (5G) communication technology, task offloading using multi-access edge computing (MEC) technologies has become an effective approach to satisfy the computation demand of various types of users. Nevertheless, most of the existing task offloading schemes overlook the unavailability of computational resources in disaster areas, failing to guarantee the quality of service (QoS) requirements of task computation in such scenarios. Therefore, this paper treats each edge node as an agent and designs a cooperative multi-agent constrained reinforcement learning (CMACRL) framework for learning the adaptive task offloading policy for each agent, ensuring the performance of emergency services and enhancing the utilization of available computational resources in disaster areas. Specifically, to satisfy the QoS requirements of applications, we formulate the constraints of applications’ performance requirements and then adopt the reward constrained policy optimization algorithm to meet the target constraints. Moreover, to mitigate the resource imbalance between disaster and safe areas, we leverage the heterogeneous-agent proximal policy optimization algorithm to generate cooperative resource use strategies for edge nodes. To evaluate the performance of CMACRL, we develop a simulator based on a real-world urban scenario. Extensive experiments are conducted, and experimental results demonstrate that CMACRL significantly improves the task offloading success rate by up to 22.49% and reduces the average service latency by up to 22.69% compared to all the baselines.
Loading