Multi-Target-Aware Dynamic Resource Scheduling for Cloud-Fog-Edge Multi-Tier Computing Network

Document Type


Publication Title

IEEE Transactions on Intelligent Transportation Systems


With the maturity of 5G and Intelligent Transportation Systems (ITS) technologies and the prospect of Beyond 5G (B5G) and 6G technologies, the limited lifetime and computing of mobile devices pose significant challenges to Quality of Service (QoS). In addition, the problem of inefficient use of computing, storage, communication, and other resources still exists in communication systems. In response to the above issues, Multi-tier Computing Networks (MTCNs) migrate computationally intensive tasks to the cloud, fog, or edge with sufficient resources, thereby realizing energy-efficient collaborative computing and multi-dimensional resource sharing. However, in the MTCN environment with complex heterogeneity, and high-intensity dynamics, how to provide sustainable solutions for resource scheduling strategies is a meaningful issue. Inspired by Virtual Network Embedding (VNE) to decouple physical network configuration, we propose a multi-target-aware dynamic resource scheduling algorithm for MTCN to improve resource flexibility, which is the first attempt in this direction. Specifically, we consider differentiated QoS requirements like computing, storage, bandwidth, delay, etc., and establish multi-target-aware embedded constraints. Additionally, we present a Deep Reinforcement Learning (DRL)-based scheduling network that can interact scientifically and efficiently with the MTCN environment. It extracts environmental information as state input to better focus on dynamic characteristics as well as calculates candidate nodes and links using a three-layer network architecture and related constraints. Furthermore, the learning process is optimized through the combination of the reward mechanism and the gradient descent mechanism. Finally, comparison experiments on three widely used evaluation indicators (long-term average revenue, long-term average revenue-cost ratio, and VNR acceptance rate) verify that the proposed algorithm has made an average improvement of $19.042\%$ , $2.563\%$ , and $3.932\%$ respectively compared with all baselines.



Publication Date



Cloud computing, deep reinforcement learning, Dynamic scheduling, Heuristic algorithms, Industrial Internet of Things, intelligent transportation systems, Multi-tier computing networks, Quality of service, resource scheduling, Search problems, sustainable solutions, Task analysis, virtual network embedding

This document is currently not available here.