iconOpen Access

ARTICLE

Heterogeneous Computing Power Scheduling Method Based on Distributed Deep Reinforcement Learning in Cloud-Edge-End Environments

Jinwei Mao1,2, Wang Luo1,2,*, Jiangtao Xu3, Daohua Zhu3, Wei Liang3, Zhechen Huang3, Bao Feng1,2, Shuang Yang1,2

1 State Grid Electric Power Research Institute Co., Ltd., Nanjing, China
2 Nanjing Nari Information & Communication Technology Co., Ltd., Nanjing, China
3 State Grid Jiangsu Electric Power Co., Ltd. Research Institute, Nanjing, China

* Corresponding Author: Wang Luo. Email: email

Computers, Materials & Continua 2026, 87(2), 85 https://doi.org/10.32604/cmc.2026.072505

Abstract

With the rapid development of power Internet of Things (IoT) scenarios such as smart factories and smart homes, numerous intelligent terminal devices and real-time interactive applications impose higher demands on computing latency and resource supply efficiency. Multi-access edge computing technology deploys cloud computing capabilities at the network edge; constructs distributed computing nodes and multi-access systems and offers infrastructure support for services with low latency and high reliability. Existing research relies on a strong assumption that the environmental state is fully observable and fails to thoroughly consider the continuous time-varying features of edge server load fluctuations, leading to insufficient adaptability of the model in a heterogeneous dynamic environment. Thus, this paper establishes a framework for end-edge collaborative task offloading based on a partially observable Markov decision-making process (POMDP) and proposes a method for end-edge collaborative task offloading in heterogeneous scenarios. It achieves time-series modeling of the historical load characteristics of edge servers and endows the agent with the ability to be aware of the load in dynamic environmental states. Moreover, by dynamically assessing the exploration value of historical trajectories in the central trajectory pool and adjusting the sample weight distribution, directional exploration and strategy optimization of high-value trajectories are realized. Experimental results indicate that the proposed method exhibits distinct advantages compared with existing methods in terms of average delay and task failure rate and also verifies the method’s robustness in a dynamic environment.

Keywords

Edge computing; end-edge collaboration; heterogeneous computing power scheduling; resource allocation

1  Introduction

With the rapid advancement of next-generation information technologies, exemplified by applications like smart factories, smart homes, and the Internet of Vehicles, the number of terminal devices worldwide has grown exponentially [1]. According to International Data Corporation, the number of global Internet of Things (IoT) device connections is projected to exceed 41 billion by 2027 [2]. This explosive growth in device scale spurs complex and diverse network access requirements: industrial robots on production floors enable real-time control via ultra-reliable low-latency communication (URLLC); smart home appliances rely on Wi-Fi to construct high-density IoT networks; and autonomous vehicles on urban roads require integration of cellular networks and roadside units for collaborative perception [3]. The data deluge from massive terminal devices presents additional challenges to the existing cloud computing paradigm. Although the cloud computing model offers powerful centralized processing capabilities, its inherent geographical distribution causes significant increases in transmission latency, failing to adequately support real-time-sensitive applications [4]. To address these issues, multi-access edge computing (MEC) has emerged as a computing paradigm that complements traditional cloud computing models [5]. MEC’s core concept is to offload computing tasks from terminal devices to the network edge via diverse access technologies, enabling flexible computing power scheduling to meet the low-latency requirements for compute-intensive applications in multi-access scenarios (e.g., smart factories, smart homes) [6]. It distributes data processing across edge servers [7], allowing users to compute near data sources via multiple access methods, thus reducing transmission latency and accelerating computation. By distributing computing resources at the network edge, MEC is better able to adapt to diverse service requirements and reduces the risk of network outages [8].

While MEC exhibits significant development potential, it faces substantial challenges in practical applications, with heterogeneous computing power scheduling being the most prominent [9]. In smart scenarios, massive data generated by terminal devices no longer needs to be uploaded to the cloud for processing-reducing the core network’s transmission burden but imposing higher demands on the resource management of edge nodes [10]. MEC nodes have limited resources yet serve a large number of terminal devices [11]; tasks must be distributed and scheduled based on their type and scale to prevent excessive loads on specific edge nodes from degrading system performance and user experience.

Significant progress has been achieved in optimizing heterogeneous computing power scheduling in MEC environments, with related studies continuously refining scheduling algorithms via deep reinforcement learning (DRL) and other methods [12]. As a specialized machine learning paradigm, DRL models systems based on Markov decision processes (MDPs) [13], enabling algorithms to adjust decisions based on environmental feedback during continuous trial-and-error-exhibiting strong performance in decision-making tasks. In MEC scenarios, DRL algorithms can dynamically adjust heterogeneous computing power scheduling strategies based on real-time edge node load, task urgency, and other key factors to achieve efficient resource utilization [14].

However, with the continuous emergence of diverse smart devices and the increasing number of terminal devices, edge networks exhibit highly dynamic environments, where edge server status (e.g., load) changes continuously over time. Against this backdrop, existing research has limitations: most algorithms assume that edge server load status can be fully acquired and updated in real time-an assumption hard to realize in practice due to monitoring overhead and privacy protection constraints. Thus, this paper leverages DRL to investigate heterogeneous computing power scheduling in MEC environments, aiming to optimize scheduling efficiency and service continuity while improving the quality of service (QoS).

2  Related Work

As the core technical support of the MEC system, the optimization level of the heterogeneous computing power scheduling strategy directly determines the quality of system services [15]. The heterogeneous computing power scheduling strategy not only needs to comprehensively consider multi-dimensional parameters such as the real-time load status of edge servers, network transmission quality, and task computing latency, but also needs to achieve a fine balance between resource utilization, cost-effectiveness, and service quality. Heterogeneous computing power scheduling enables edge servers and users to focus on their respective core businesses, thereby maximizing resource utilization, reducing costs, and enhancing overall service performance [16].

In recent years, machine learning techniques have shown significant advantages in the field of load optimization for complex systems. Especially in the edge computing scenario, the task offloading decision faces a highly dynamic and heterogeneous complex system environment, involving the contradiction between multi-dimensional resource constraints (of terminal devices, communication links, and edge nodes) and real-time requirements, which poses a modeling bottleneck for meta-heuristic algorithms. In this context, DRL can simultaneously model the relationships among network bandwidth fluctuations, server load, and task priority by constructing an end-to-end “state-action-reward” decision-making optimization framework, and its online learning characteristics provide a new technical path for adaptive task offloading in complex edge environments.

Zhang et al. [9] proposed a heterogeneous task collaborative offloading scheme for satellite edge computing, which effectively reduces satellite energy consumption and improves task processing efficiency through a ground-satellite-cloud multi-layer architecture and dynamic threshold strategy. Gao et al. [17] proposed a DRL-based algorithm with continuous convex approximation for hybrid UAV-assisted systems, optimizing service assignment, task segmentation, UAV trajectories, resources, and transmission power to reduce energy consumption. Cai et al. [18] designed a framework based on the Double Deep Q-Network (DDQN), along with a resource tree model and multi-granular task decomposition, for cloud-edge-device networks, aiming to optimize task decomposition, hierarchical scheduling, and maximize latency-energy utility. Chen et al. [19] developed an online convolutional DRL algorithm for Non-Orthogonal Multiple Access (NOMA) enabled edge networks, decoupling the minimization of system computing completion time into duration and resource allocation sub-problems. Fan et al. [20] presented a dual-delay deep deterministic policy gradient algorithm for in-vehicle edge computing, using road side unit (RSU) collaboration to optimize task offloading and resource allocation, cutting down delay and energy use. Guo et al. [21] put forward a fuzzy trust-enhanced, DRL-driven scheme for vehicle-mounted edge networks, optimizing offloading delay and task processing credibility. Li et al. [22] built a pre-determined multi-agent DRL offloading mechanism for resource-constrained edges to maximize user experience quality. Li et al. [23] proposed a Lyapunov- and multi-agent deep deterministic policy gradient-based two-stage offloading architecture for distributed environments, minimizing subsystem delay through joint initial and RSU peer-to-peer offloading while considering RSU energy constraints. Lin et al. [24] formalized decentralized edge task offloading as partially observable Markov decision-making process (POMDP), leveraging heuristic logs to warm-up and fine-tune a DRL model, enhancing task success and solving cold-start issues. Liu et al. [25] constructed a two-time-scale multi-dimensional resource optimization model with a dual-actor deep deterministic strategy gradient for edge networks, reducing long-term delay and cache costs. Liu et al. [26] proposed a predictive Deep Q-Network (DQN) for resource allocation and computation offloading (PQ-RACO) algorithm and a multi-agent DQN for resource allocation and computation offloading (MQ-RACO) algorithm for the dynamic edge networks, which optimize offloading decisions and power allocation via agent behavior prediction and explicit reward design to maximize computing rates.

When applying DRL to construct a task offloading model, the existing research fails to effectively model the dynamic characteristics in large-scale terminal device scenarios:

(1) When the number of terminal devices surges, the load fluctuation of edge servers, the instantaneous change of network link state, and the heterogeneity of computing resources will lead to the high dynamic time variation of the system, which significantly reduces the stability of the state transition probability in the Markov Decision Process (MDP) based framework and makes it challenging to ensure the strategy convergence.

(2) The state space dimensional explosion and the complexity of action space combination caused by massive device access have forced the algorithm to frequently retrain the model and update the policy parameters, which not only causes the exponential increase of computing overhead, but also aggravates the system’s dynamic imbalance due to the real-time response lag. The over-reliance on the dynamic Markov assumption and the scalability defect of the high-dimensional state-action space seriously restrict the practical value of the algorithm in ultra-large-scale edge computing scenarios.

3  System Model

This section abstracts the MEC system into three layers, as illustrated in Fig. 1. The first layer is a third-party trusted authority, geographically remote from terminal devices, which primarily stores task offloading-related terminal device information. The second layer comprises edge nodes, including edge servers, base stations, and other edge network access points, where edge servers are equipped with sufficient computing resources; base stations are responsible for direct communication with terminal devices within their coverage area, handling wireless signal transmission and reception; and edge network access points integrate fixed-network-connected devices into the edge network. The third layer consists of a large number of terminal devices-typically resource-constrained devices enabled by the Industrial IoT and smart city initiatives. As shown in the figure, devices in the MEC environment include IIoT terminal devices and mobile robots connected via 5G, smart home devices connected via Wi-Fi access points, environmental monitoring sensors in smart cities connected via wired networks, and other smart devices using fixed broadband or Wi-Fi.

images

Figure 1: An overall cloud-edge-end collaboration task offloading scenario.

This section focuses on a MEC environment involving a set of edge servers denoted as E={1,2,,e} and a set of terminal devices denoted as D={1,2,,d}. It specifically addresses task offloading over a set of time slots T={1,,t}, where each time slot has a short duration δ (in seconds).

3.1 Task Model

Assume a new task arrives at a terminal device at the start of a time slot. Let Id(t) denote the unique index of the task arriving at the d-th terminal device during time slot t. For generality, Id(t)=0 if no new task arrives at terminal d in that time slot.

Let ηd(t) denote the data size (in bits) of newly arrived task Id(t), with ηd(t)=0 if no new task arrives at terminal d. Tasks have distinct processing densities ρd,t (CPU cycles per bit). Each task also has a latency constraint τd,t (in time slots): task Id(t) is discarded if not fully processed by the end of time slot t+τd,t1.

3.2 Transmission Model

Edge servers use non-overlapping frequency bands, eliminating uplink interference between them. Let |hd,e|2 represent the channel gain from terminal d to edge server e, and pd,e denote terminal d’s transmission power when communicating with edge server e.

The transmission rate from terminal d to edge server e is Od,etran, given by:

Od,etran=Blog2(1+|hd,e|2pd,eσ2),dD,eE(1)

where B is the channel bandwidth and σ2 is the channel noise power.

At the start of a time slot, if the local computing capacity of terminal d is insufficient for task Id(t), Tdtran(t) denotes the time slot when the task is transmitted (to an edge server) or discarded (if offloading fails). If the task is processed locally or no task arrives, Tdtran(t)=0 (with Tdtran(0)=0).

Let μdtran(t) (in time slots) represent the waiting time for transmission of task Id(t), determined by prior transmission slots Tdtran(t) (where t<t) as:

μdtran(t)=max{supt{0,1,,t1}Tdtran(t)t+1,0}(2)

If task Id(t) starts waiting for transmission in a time slot, Tdtran(t) (the time slot when transmission completes or the task is discarded) is calculated as:

Tdtran(t)=min{t+μdtran(t)+eEyd,e(t)ηd(t)δOd,etran1,t+τd,t1}(3)

where yd,e(t)=1 indicates that task Id(t) is offloaded to edge server e, and yd,e(t)=0 otherwise.

3.3 Computational Models

When a new task Id(t) arrives at terminal d in a time slot, the terminal must decide whether to process the task locally or offload it to an edge server.

The indicator xd(t) specifies whether task Id(t) is processed locally (xd(t)=1) or offloaded to the edge (xd(t)=0). Here, ηd(t)xd(t) represents the data volume processed locally, and ηd(t)(1xd(t)) represents the data volume transmitted to edge servers.

If task Id(t) is offloaded to edge server e, yd,e(t)=1 (0 otherwise). Each task can only be offloaded to one edge server, so:

eEyd,e(t)={1if xd(t)=0,0if xd(t)0.(4)

3.3.1 Local Computing

Let fd denote the computing capacity of terminal d. For time slot t, Tdcomp(t) denotes the time slot when task Id(t) is processed locally or discarded; Tdcomp(t)=0 if the task is not processed locally (e.g., Id(t)=0).

Let μdcomp(t) represent the waiting time for local processing of the task Id(t), determined by prior local processing slots Tdcomp(t) (where t<t) as:

μdcomp(t)=max{supt{0,1,,t1}Tdcomp(t)t+1,0}(5)

If task Id(t) starts waiting for local processing in a time slot, Tdcomp(t) (the time slot when processing completes or the task is discarded) is:

Tdcomp(t)=min{t+μdcomp(t)+ηd(t)ρd,tδfd1,t+τd,t1}(6)

3.3.2 Task Offloading

Each edge server e maintains a computation queue for each terminal dD. Let Id,e(t) be the unique index of the task from d added to e’s queue at the start of time slot t (Id,e(t)=0 if no such task exists). Let ηd,e(t) (in bits) be the task’s data size (ηd,e(t)=0 if no task arrives).

Edge server computation queues follow First-In-First-Out (FIFO). Let qd,e(t) (in bits) be the length of the queue of d in e at the end of time slot t. Let qe(t) denote the set of non-empty queues in e at time slot t:

qe(t)={dηd,e(t)>0qd,e(t1)>0}(7)

Let Qe(t)=|qe(t)| be the count of non-empty queues in e in time slot t.

Let fe (in CPU cycles per second) be computing capacity of edge server e. Let dropd,e(t) (in bits) be the data dropped from d’s queue on e at the end of the time slot t. The queue length qd,e(t) updates as:

qd,e(t)={(qd,e(t1)+ηd,e(t))feδρd,tQe(t)(1(dqe(t))dropd,e(t)),1(dqe(t))dropd,e(t)0,0,1(dqe(t))dropd,e(t)<0,(8)

where 1() is the indicator function (1 if the condition holds, 0 otherwise).

For a terminal task d added to edge server e’s queue at the start of the time slot t, Td,ecomp(t) denotes the time slot when the task is processed or discarded; Td,ecomp(t)=0 if Id,e(t)=0.

Let T^d,ecomp(t) denote the time slot when processing starts:

T^d,ecomp(t)=max{t,supt{0,1,,t1}Td,ecomp(t)+1}(9)

The task size ηd,e(t) satisfies: it is not larger than the total processing capacity of edge server e from T^d,ecomp(t) to Td,ecomp(t), but greater than the total capacity from T^d,ecomp(t) to Td,ecomp(t)1. Formally:

t=T^d,ecomp(t)Td,ecomp(t)δfeρd,tQe(t)1(dqe(t))ηd,e(t)(10)

t=T^d,ecomp(t)Td,ecomp(t)1δfeρd,tQe(t)1(dqe(t))<ηd,e(t)(11)

3.3.3 Task Computation Latency

Task computation latency is defined as the sum of waiting time and processing time. Let delayd(t) (in time slots) denote the latency of the task Id(t).

If processed locally:

delayd(t)=Tdcomp(t)t+1(12)

If offloaded to an edge server:

delayd(t)=eEt=tT1(Id,e(t)=Id(t))Td,ecomp(t)t+1(13)

3.4 Problem Definition

The optimization objective is to minimize the task computation latency: given the current system state, determine the optimal edge server for offloading such that task Id(t)’s latency delayd(t) satisfies delayd(t)τd,t.

The MEC environment exhibits continuous time-varying characteristics, and traditional algorithms lack scalability in large-scale dynamic scenarios. Thus, we employ DRL to address this problem. Formally, the terminal device policy is a mapping from system states to actions, denoted as πd:𝒮𝒜. Defining task computation latency as the offloading cost, minimizing this cost corresponds to minimizing latency. Let γ(0,1] be the discount factor (weighting future costs). The goal is to find each terminal’s optimal policy to minimize the expected long-term latency:

πd=argminπdE[tTγt1delayd(t)πd](14)

4  Heterogeneous Computing Power Scheduling Methods

The MEC environment is dynamically changing, and the load uncertainty of numerous terminal devices and edge servers renders traditional single-agent reinforcement learning ineffective for task offloading decision-making, even failing to converge in extreme cases. Thus, this section proposes a multi-agent reinforcement learning algorithm, centralized replay buffer-based deep recurrent Q-network (CRB-DRQN), enabling terminal devices and edge servers to make offloading decisions that satisfy the requirements of delay-sensitive tasks under unknown load conditions. Fig. 2 illustrates the overall execution process of the algorithm, involving three types of actors: a third-party trusted authority responsible for the collection and storage of experience, edge servers for network training and task computation, and terminal devices with offloading decision capabilities. The system architecture adopts a fusion mode that enables cloud-based training and edge-based execution, laying a theoretical foundation for large-scale deployment. A single edge server manages multiple terminal devices. Edge servers are responsible for training neural networks but do not directly participate in task offloading decision-making. Instead, offloading decisions are determined and executed by terminal devices. Compared with centralized algorithms, when edge servers have sufficient computing resources, the proposed architecture can support a larger number of terminal devices. When the number of terminal devices is enormous, the computing capacity can be linearly expanded by increasing the number of edge servers.

images

Figure 2: The overall execution process of CRB-DRQN.

To reduce the impact of changes in the MEC environment on agent training, this section uses an ensemble of all agents’ experience trajectories to train the neural network of each agent. However, the agent’s experience trajectory contains private data such as task information. Therefore, this section builds a third-party trusted authority in the cloud. Each agent uploads its own experience trajectory to the third-party trusted authority, which is responsible for integrating the experience trajectories to form a global experience trajectory, storing it in the central trajectory pool, and then training the corresponding neural network.

When a new task arrives, the terminal device requests the neural network weights of the Evaluation Q-network from the corresponding edge server. Based on the offloading decision generated by the neural network, the terminal device offloads the task to the designated edge server, and the edge server returns the task’s calculation result to the terminal device. The terminal device collects experience samples and stores them in the experience replay buffer. When the number of experience samples stored in the experience replay buffer reaches a preset threshold, they form an experience trajectory. The terminal device then uploads this experience trajectory to the central trajectory pool of the third-party trusted authority, which integrates all uploaded experience trajectories. The edge server adopts similarity-based sampling to extract experience trajectories from the central trajectory pool, uses them to train the agent, and updates the network parameters accordingly.

4.1 POMDP Build

4.1.1 Status

This section assumes each edge server broadcasts its number of non-empty computation queues at the end of each time slot. The matrix Load(t) represents each edge server’s historical load levels over the previous C time slots. Let Qj(t) denote the load level of edge server j at time t; Load(t) is a C×E matrix with elements defined as:

{Load(t)}i,j=Qj(tC+i1)(15)

At the beginning of time slot t, the terminal device d observes the state:

sd(t)=(ηd(t),μdcomp(t),μdtran(t),qd,e(t1),Load(t))(16)

4.1.2 Action

At the start of time slot t, if a new task arrives at terminal device d, d selects an action for this task. The action of terminal device d in time slot t is denoted as:

ad(t)=(xd(t),yd(t))(17)

Let Space denote the action space, where Space={0,1}1+E.

4.1.3 Rewards

This section uses negative rewards associated with task delays. If task Id(t) has been processed, the reward is:

Rd(t)=delayd(sd(t),ad(t))(18)

On the other hand, if task Id(t) has been discarded, then:

Rd(t)=Const(19)

where Const<0 is a negative constant.

4.2 Neural Network Structure

As shown in Fig. 2, the Target and Evaluation Q-networks share the same structure on edge servers. The Target Q-network stabilizes learning and computes target Q-values; the Evaluation Q-network handles policy evaluation and parameter updates.

The Evaluation Q-network calculates Q-values for all actions from the current state and selects the highest-value action for execution. This guides the agent’s decision-making, enabling near-optimal action selection, interaction with the environment, and gradual learning of the optimal policy.

The structure of the Evaluation Q-network on terminal devices matches that on edge servers. The core goal of the neural network is to map each state to a set of Q-values corresponding to possible actions. This mapping process involves the collaborative work of multiple network layers.

State information is passed through the input layer to the neural network. The state information includes ηd(t),μdcomp(t),μdtran(t),qd,e(t1),Load(t), where ηd(t),μdcomp(t),μdtran(t),qd,e(t1) are passed to the fully connected layer, and Load(t) (for edge server load level prediction) is passed to the long short-term memory (LSTM) layer.

The LSTM layer learns the dynamics of edge server load levels and predicts near-future load levels using a LSTM unit. Hidden neurons in the LSTM are sequentially connected to matrix rows to track sequence changes (e.g., load level variations between time slots). Outputs from the last LSTM cell (indicating future load dynamics) are passed to the fully connected layer as Load(t)c. Two such layers use ReLU-activated neurons, with connections between anterior/posterior layers for information transmission.

The Advantage-Value layer includes Advantage Network (A fully connected network learning action advantage values A(sd(t),a;θa), where θa denotes network weights) and Value Network (A fully connected network learning state values V(sd(t);θv), where θv denotes network weights).

These values are computed through neural network layers (input Advantage-Value) and trained on the edge server. The output layer calculates the Q-value of state-action pairs via:

Q(sd(t),a)=V(sd(t);θv)+[A(sd(t),a;θa)1|Space|aSpaceA(sd(t),a;θa)](20)

Here, the second term adjusts the advantage value relative to the average advantage of all actions in the state.

4.3 Familiarity Trajectory Sampling

4.3.1 Trajectory

This section defines an experience sample obtained by the terminal device at time t as: (sd(t),ad(t),rd(t),sd(t+1)). This tuple includes: Current state sd(t), Action ad(t) taken,Reward rd(t) received, Next state sd(t+1). It is stored in the replay buffer local to the terminal device.

Traditional DQN stores single-step experiences in a replay buffer. This section proposes a familiarity trajectory sampling algorithm that uses temporal sequences of multiple experiences as the agent’s trajectory. When a terminal device stores an experience sample, it shares the trajectory starting at time t to a central trajectory pool (managed by a third-party entity). A trajectory is defined as:

Tra={(sd(t),ad(t),rd(t),sd(t+1)),(sd(t+1),ad(t+1),rd(t+1),sd(t+2)),(sd(t+n),ad(t+n),rd(t+n),sd(t+n+1))}(21)

where n is the number of multi-step experiences (a constant).

Let 𝒯={Tra1,Tra2,} denote the set of all trajectories in the central pool. The central pool organizes trajectories uploaded by terminal devices into a replay buffer, from which edge servers sample during training. The size of 𝒯 is denoted |𝒯|.

4.3.2 Familiarity Trajectory Sampling

Based on prior random sampling methods, this section proposes a familiarity trajectory sampling strategy. The core idea is: trajectories with fewer historical samples should have a higher re-sampling probability. Define Ni as the familiarity of trajectory Trai (increases with each sampling). As Ni grows, the probability of re-sampling decreases.

The sampling probability of trajectory Trai is:

P(i)=(ln(Ni)Ni)α2p=1|𝒯|(ln(Np)Np)α2(22)

where |𝒯| denotes the number of trajectories in the central pool, α is a hyperparameter that balances the importance of priority, Ni represents the familiarity (sampling count) of Trai, and P(i) is the sampling probability of Trai.

4.4 Task Offloading Algorithm

First, initialize the Target Q-network, Evaluation Q-network, and counter counte of the edge server e; for the terminal device d, initialize its initial state sd(1) and trajectory counter nd, and download the network weights θd of the Evaluation Q-network corresponding to the edge server e. When the time slot tT starts, if a new task Id(t) arrives at the terminal device d, the terminal device d will select an action ad(t) according to the ε-greedy strategy and observe the state information sd(t+1) of the next time slot t+1. Meanwhile, it judges according to the trajectory storage condition: if the trajectory counter nd satisfies ndmodn=0 (where n is the trajectory length), the current trajectory batch Trad,i is sent to the third-party trusted authority and the trajectory index i is incremented; otherwise, the state transition experience (sd(t),ad(t),rd(t),sd(t+1)) is appended to the local trajectory and the trajectory counter nd is updated. In addition, if the current time slot t meets the periodic network weight download condition (mod (t,tag)==0), the terminal device d will download the updated Evaluation Q-network weights θe from the associated edge server e. As shown in Fig. 3a and Algorithm 1.

images

Figure 3: Flowchart of the CRB-DRQN Algorithm. Figure (a) illustrates the execution flow of the CRB-DRQN algorithm on terminal devices. Figure (b) illustrates the execution flow of the CRB-DRQN algorithm on edge devices.

Meanwhile, the edge server e extracts the trajectory set I from the central trajectory pool based on the familiarity sampling mechanism, updates the familiarity of each trajectory. For each experience sample Ij,k (the k-th experience sample of the j-th trajectory in the trajectory set I) in each trajectory TrajI, the target Q-value Q^e,Ij,kTarget is calculated using Eq. (23). As shown in Fig. 3b and Algorithm 2.

Q^e,Ij,kTarget=Rj,k+γQe(sj,k+1,aj,kNext,θe)(23)

The target Q-value Q^n,Ij,kTarget reveals the expected long-term reward of the action aIj,kNext given the state sIj,k+1, i.e., the sum of the reward in the experience sample Ij,k and the Q-value of the action that may be selected under the given conditions. Among them, aIj,kNext can be calculated using Eq. (24).

aj,kNext=argminaSpaceQe(sj,k+1,a,θe)(24)

Subsequently, the gradient descent method is used to minimize the loss function L(θe,Q^eTarget) and update the Evaluation Q-network parameters θe via Eq. (25), while incrementing its training counter counte. When counte satisfies countemodupdate==0, the edge server synchronizes the Evaluation Q-network parameters θe to the Target Q-network parameters θe, completing the delayed update of the network parameters.

L(θe,Q^eTarget)=1n|I|Ij,kI(Q^e,Ij,kTargetQe(sIj,k+1,aIj,k,θe))2(25)

images

The time complexity of this algorithm is O(T×(|D|+|E|×I×K)), where T denotes the total number of time steps, |D| and |E| represent the number of terminal devices and edge servers, respectively, I is the number of trajectories sampled in a single iteration, and K denotes the number of experience samples per trajectory. Its core consists of an initialization phase and a time-step main loop: The initialization requires traversing all edge servers O(|E|) and terminal devices O(|D|) to complete network parameter loading; in the main loop, each time step involves processing task action selection and trajectory recording for all terminal devices O(|D|), as well as trajectory sampling and experience calculation O(I×K) for all edge servers O(|E|). The space complexity is O(|E|+|D|), which stems from the independent storage of Q-network parameters and counters by edge servers and terminal devices, with no dependence on task scale. The algorithm enables parallel processing of terminal devices and edge servers, reducing synchronization overhead and adapting to the asynchronous decision-making requirements of end-edge collaboration.

images

4.5 Convergence Analysis

Based on the theory of Bellman operator, this section conducts a rigorous theoretical analysis on the convergence of the CRB-DRQN algorithm in the heterogeneous cloud-edge-end environment. By transforming the POMDP into a belief MDP, defining the corresponding Bellman operator and proving its contractivity, the convergence characteristics of the algorithm are finally derived based on the Banach fixed-point theorem.

Let the true state of the system be s𝒮, the observation sequence of terminal device d at time slot t be o1:t, and the action sequence be a1:t1. The belief state b is defined as the probability distribution of the true state:

b(s)=P(s|o1:t,a1:t1)(26)

The belief state space is the set of all possible probability distributions. Based on the belief state, a belief MDP can be constructed as a 5-tuple (,𝒜d,𝒯b,b,γ), where 𝒜d denotes the action space, 𝒯b is the state transition function, b represents the reward function, and discount factor γ is 0.9.

Under the framework of belief MDP, the policy of terminal device d is defined as πd:𝒜d. The state-value function and action-value function corresponding to policy πd are respectively given by:

Vπd(b)=Eπd[k=0γkrt+k|bt=b](27)

Qπd(b,a)=Eπd[rt+γVπd(bt+1)|bt=b,at=a](28)

For the double Q-network architecture of the CRB-DRQN algorithm, define the Bellman expectation operator for the Evaluation Q-network as:

TπdQ(b,a)=b(b,a)+γEo,b𝒯b[Q(b,πd(b))](29)

Define the Bellman optimality operator for the Target Q-network as follows:

TQ(b,a)=b(b,a)+γEo,b𝒯b[maxa𝒜dQ(b,a)](30)

To prove the convergence of the algorithm, it is necessary to verify that the aforementioned Bellman operator is a contraction mapping. Define the action-value function space Q as the set of all bounded functions, and its infinity norm is given by:

Q1Q2=maxb,a𝒜d|Q1(b,a)Q2(b,a)|(31)

Theorem 1: The Bellman expectation operator 𝒯πd is a γ-contraction mapping.

Proof: For any Q1,Q2𝒬, the following holds:

|𝒯πdQ1(b,a)𝒯πdQ2(b,a)|=γ|E[Q1(b,πd(b))Q2(b,πd(b))]|γE[|Q1(b,πd(b))Q2(b,πd(b))|]γQ1Q2(32)

Taking the supremum yields 𝒯πdQ1𝒯πdQ2γQ1Q2, and the proof is completed.   □

Theorem 2: The Bellman optimality operator 𝒯 is a γ-contraction mapping.

Proof: For any Q1,Q2𝒬, the following holds:

|𝒯Q1(b,a)𝒯Q2(b,a)|=γ|E[maxaQ1(b,a)maxaQ2(b,a)]]γE[maxa|Q1(b,a)Q2(b,a)|]γQ1Q2(33)

Thus, 𝒯 is also a γ-contraction mapping.

According to the Banach fixed-point theorem, a contraction mapping has a unique fixed point in a complete metric space. Therefore, the Bellman expectation operator 𝒯πd and the optimality operator 𝒯 have unique fixed points Qπd and Q, respectively, which satisfy:

TπdQπd=Qπd,TQ=Q(34)

Let the Q-function at the k-th iteration be Qk, then the iteration error satisfies:

QkQγkQ0Q+ε1γ(35)

where ε denotes the function approximation error. As k, γk0, and the algorithm converges.    □

4.6 Discussion on Practical Issues

First, in the task offloading problem, traditional methods often rely on global state information, resulting in high communication costs. This paper adopts a partially observable modeling approach, enabling terminal devices to make decisions based solely on local observations. This design reduces the frequency of state transmission.

Second, the terminal devices in this study only need to obtain minimal local state information, reducing the risk of privacy leakage from the source.

Finally, the system architecture of this paper adopts a “centralized training-distributed execution” mode, which provides a theoretical basis for large-scale deployment. This modular design allows the system to linearly expand its processing capacity by adding edge nodes.

5  Performance Evaluation

This section details the experimental setup and results of the algorithm proposed in this section in a simulation environment. The evaluation metrics include average task delay, task failure rate under varying terminal device counts and task arrival rates, as well as the reward curve of each algorithm. Simulations are conducted on a desktop equipped with an AMD Ryzen 7 5700X 8-Core Processor, an NVIDIA GeForce RTX 3070Ti GPU, and 32 GB RAM.

5.1 Experimental Parameter Settings

A simulation scenario with four edge servers is built. Specific parameters are in Table 1. Neural network settings: batch size = 4, learning rate = 0.001, discount factor = 0.9, random exploration probability decays from 1 to 0.01. The root mean square propagation (RMSPROP) optimizer is used. The configuration of experimental parameters is based on the typical characteristics of real-world edge computing environments and research requirements. First, the time slot duration δ is set to 100 milliseconds to match the fine-grained time unit for task scheduling in edge computing, thereby avoiding delay inaccuracies caused by excessive discretization. The CPU frequency of terminal devices (fd = 2.5 GHz) and the frequency of edge servers (fe = 41.8 GHz) refer to the common configurations in [27], which reflects the heterogeneity of resources. The transmission rate Od,etran = 14 Mbps corresponds to the bandwidth of a typical wireless network, while the processing density ρd,t = 500 is determined based on the average number of CPU cycles required per unit of data, so as to balance the computational load. The delay constraint τd,t = 12 time slots is tailored for delay-sensitive applications, ensuring that tasks are completed within a reasonable time. Overall, the selection of parameters aims to simulate dynamic multi-access scenarios, verify the robustness of the algorithm under resource constraints, and provide a consistent benchmark for comparative experiments.

images

5.2 Introduction to Experimental Comparison Methods

The proposed task offloading algorithm (CRB-DRQN) is compared with the following benchmark algorithms:

(1) Decentralized algorithm: A decentralized actor-critic algorithm which contrasts with the centralized training/distributed execution framework of the proposed algorithm. Agents update networks independently using local observations, with no inter-agent information sharing.

(2) Local-Only strategy: Forces all agents to process computational tasks locally, omitting any offloading decision-making.

(3) Q-value Mixing (QMIX) algorithm [28]: A value-learning-based multi-agent reinforcement learning algorithm for solving cooperative task interaction/reward distribution challenges. It uses cross-entropy optimization for action selection, which is adapted to continuous action spaces. Implementation follows literature [29].

(4) Multi-Agent Proximal Policy Optimization (MAPPO) algorithm [30]: An actor-critic algorithm for multi-agent scenarios, optimized for continuous action spaces and massively parallel computing environments.

(5) Joint Optimal DRL Algorithm with Privacy Preservation (JODRL-PP) [5]: An actor-critic-based task offloading algorithm. It addresses incomplete multi-user information sharing via a third-party centralized training network, enabling effective offloading strategy learning.

(6) Multi-Tier offloading model [31] based on game theory principles. It addresses the limitations of edge server computational resources and inefficient resource allocation in vehicular networks by dividing the offloading process into two stages.

5.3 Evaluation of Experimental Results

5.3.1 Convergence of the Algorithm

Fig. 4 intuitively demonstrates the fast convergence property of the CRB-DRQN algorithm by comparing the average reward evolution curves of different algorithms. Experimental data show that the CRB-DRQN algorithm reaches steady-state convergence after approximately 170 training episodes, with the final average reward stabilizing in the −0.5 range-significantly outperforming the other benchmark algorithms (e.g., the JODRL-PP algorithm requires 410 episodes to converge and exhibits reward jump phenomena). This efficient convergence stems from the core design of the algorithm: first, the Double DQN and Target network mechanism effectively suppress the overestimation of Q-values by decoupling action selection and value evaluation; second, the centralized training architecture integrates global experience trajectories through a third-party trusted authority, enabling agents to obtain an approximate global perspective in partially observable environments and avoiding policy oscillations caused by environmental non-stationarity in decentralized algorithms. In addition, the similarity-based trajectory sampling mechanism further optimizes exploration efficiency by dynamically adjusting the sampling probability of high-value experiences, promoting the smooth approximation of the Q-function to the optimal solution.

images

Figure 4: Average rewards of different algorithms and Convergence time of CRB-DRQN algorithm.

Fig. 4 also illustrates the variation trend of the convergence time of single agent under different terminal device scales, and its linear growth pattern verifies the algorithm’s stability in dynamic scaling scenarios. When the CRB-DRQN algorithm faces the expansion of large-scale terminal devices and edge server clusters, its distributed execution architecture ensures system scalability through a three-layer mechanism. First, the “centralized training-distributed execution” framework enables terminal devices to make independent real-time decisions based on local observed states, while edge servers achieve dynamic allocation of computational loads through asynchronous training and parameter updates, effectively avoiding single-point bottlenecks. Specifically, when the number of terminal devices increases from 5 to 35, the convergence time only rises approximately linearly, indicating that the algorithm has strong robustness to scale expansion. When the algorithm is scaled to larger-scale scenarios, the computational pressure can be dispersed by adding edge nodes, and the processing capacity can be expanded linearly.

5.3.2 Comparison of Task Delay and Failure Rate under Different Terminal Device Scales

Fig. 5a shows the comparison results of the average latency of each algorithm task at different terminal device scales. Although the average latency of all algorithms is controlled within the latency constraints, it should be noted that this metric only counts successfully processed tasks and does not include discarded tasks due to insufficient resources. Except for the Local-Only algorithm, the average computing latency of the other algorithms increases monotonically with the expansion of the terminal device. Compute latency is related to the computing resources and load of the device. As the number of terminal devices increases, the number of tasks that each edge server needs to process increase, which directly leads to an increase in queue wait times, resulting in an increase in the average latency of computing. The CRB-DRQN algorithm maintains a low latency level under different terminal device scales. When the number of terminal devices is 5, the latency (0.50 s) is 46.8%, 15.3%, and 13.8% lower than that of the Decentralized algorithm (0.94 s), the MAPPO algorithm (0.59 s), and the JODRL-PP algorithm (0.58 s), respectively, slightly higher than the QMIX algorithm (0.48 s). It is worth noting that the latency gap between the JODRL-PP algorithm and the CRB-DRQN algorithm gradually decreases with the increase of device scale, and when the number of terminal devices reaches 25, the latency value of the two algorithms tends to be the same level (CRB-DRQN algorithm: 0.762 s, JODRL-PP algorithm: 0.767 s). This phenomenon can be attributed to the performance optimization of the CRB-DRQN algorithm through the following mechanisms: firstly, by improving the neural network, it has better load awareness ability in dynamically changing scenarios, and secondly, through the global experience sharing mechanism realized by the central trajectory pool, so that each agent can learn the optimal offloading strategy collaboratively, effectively reducing the local decision bias.

images

Figure 5: Task delay and failure rate under varying scales of terminal devices. Figure (a) shows the comparison results of the average latency of each algorithm task at different terminal device scales. Figure (b) presents how terminal device scale affects task failure rates.

Fig. 5b presents how terminal device scale affects task failure rates. Experimental data indicate that at 25 devices, task failure rates diverge significantly: Local-Only fails to handle high concurrency due to limited local resources. The Decentralized and QMIX algorithms see linear failure rate growth with scale expansion, reaching 0.42 and 0.37 respectively at 25 devices. MAPPO shows a conservative strategy, with its failure rate stably within 0.24 ± 0.02. Conversely, JODRL-PP and Multi-tier offloading control the failure rate at 0.04 for 25 devices. The proposed CRB-DRQN algorithm achieves global state awareness via the central trajectory pool, stabilizing the final task failure rate at 0.017%–57.5% lower than JODRL-PP and the Multi-tier offloading strategy. This validates the empirical playback mechanism’s role in enhancing system reliability.

5.3.3 Comparison of Task Delay and Failure Rate under Different Task Scales

This section compares the convergence of the algorithm, the task delay and failure rate at different terminal device scales, and the task delay and failure rate at different task scales.

Fig. 6a presents the average latency across different task sizes. As the task arrival rate rises from 1.0 to 3.0 Mbps, all algorithms’ average latencies increase. At 1.0 Mbps, our CRB-DRQN algorithm (0.43 s) outperforms Decentralized (0.90 s), QMIX (0.53 s), and MAPPO (0.71 s), matching JODRL-PP (0.43 s) and Multi-tier offloading (0.42 s) but slightly higher than Local-Only (0.38 s). When the rate hits 3.0 Mbps, Local-Only latency spikes to 0.89 s (and keeps rising) due to terminal devices’ local computing resource constraints. Notably, Local-Only still outperforms Decentralized, QMIX, and MAPPO, whose longer waiting times create efficiency bottlenecks in their offloading strategies. CRB-DRQN maintains the lowest latency between 1.5–2.5 Mbps. At 3.0 Mbps, its 0.73 s latency is marginally higher than JODRL-PP’s 0.72 s and Multi-tier offloading 0.72 s, revealing CRB-DRQN’s optimization boundary under high load.

images

Figure 6: Task delay and failure rate under varying task scales. Figure (a) presents the average latency across different task sizes. Figure (b) compares task failure rates under different task sizes.

Fig. 6b compares task failure rates under different task sizes. Experimental data indicate that as task arrival rate rises from 1.0 to 3.0 Mbps, failure rates of the Decentralized and Local-Only algorithms soar from 0.18, 0.014 to 0.52, 0.37, respectively. The former suffers from environmental instability, hampering effective strategy execution; the latter is limited by insufficient local computing power to meet demands. The JODRL-PP, Multi-tier offloading, and CRB-DRQN algorithms exhibit strong reliability and robustness, with maximum failure rates of 0.052 (JODRL-PP, 3.0 Mbps), 0.051 (Multi-tier offloading, 3.0 Mbps) and 0.057 (CRB-DRQN, 3.0 Mbps). Within 1.0–2.5 Mbps task arrival rate, CRB-DRQN outperforms others; yet at 3.0 Mbps, its failure rate (0.057) marginally exceeds JODRL-PP’s 0.052 and Multi-tier offloading’s 0.056. Fig. 6 suggests CRB-DRQN still has optimization potential under high load, possibly due to mismatches between neural network-perceived and actual load levels.

6  Conclusion

To address core challenges in MEC environments-including a large number of terminal devices and frequent dynamic changes in system states-this section proposes a end-edge collaborative task offloading method based on POMDP. A three-layer offloading framework encompassing edge computing resources, terminal devices, and third-party trusted authority is constructed to solve the policy mismatch problem caused by environmental dynamics in traditional algorithms. At the modeling level, a slotted transmission and computation model is innovatively introduced to quantify task processing delay and queue dynamics, and a partially observable state space is built based on edge server load history to reduce system information interaction overhead.

The proposed CRB-DRQN algorithm adopts a hybrid architecture of centralized experience sharing and distributed policy execution: through a central trajectory pool constructed by third-party trusted authority, it achieves safe integration and efficient reuse of cross-agent experience trajectories; a familiarity trajectory sampling mechanism is designed to optimize the distribution of training samples, effectively alleviating the policy lag problem of traditional experience replay mechanisms in dynamic environments; the neural network structure designed based on DDQN and LSTM enhances the prediction ability of edge server load fluctuations, enabling agents to make accurate offloading decisions under partially observable conditions.

Experimental results show that CRB-DRQN exhibits significant advantages in scenarios with expanded terminal device scales and dynamic changes in task loads, providing a new solution for task offloading in MEC environments and effectively improving system reliability and offloading efficiency.

In real-world environments, the dynamic changes of tasks and resources are often more frequent and complex. The performance of the algorithm can be evaluated in a demonstration system, thereby identifying and resolving more potential issues. Additionally, game theory can be combined with multi-agent reinforcement learning techniques to further explore the mechanism of strategic interaction among multiple devices.

Acknowledgement: Not applicable.

Funding Statement: This work was funded by the State Grid Corporation Science and Technology Project “Research and Application of Key Technologies for Integrated Sensing and Computing for Intelligent Operation of Power Grid”(Grant No. 5700-202318596A-3-2-ZN).

Author Contributions: The authors confirm contribution to the paper as follows: Conceptualization, Wang Luo and Bao Feng; methodology, Wang Luo and Jinwei Mao; software, Jinwei Mao, Shuang Yang, Daohua Zhu and Wei Liang; validation, Shuang Yang, Jiangtao Xu and Zhechen Huang; formal analysis, Jinwei Mao and Bao Feng; investigation, Shuang Yang and Jiangtao Xu; resources, Daohua Zhu and Wang Luo; data curation, Jiangtao Xu, Zhechen Huang and Wei Liang; writing—original draft preparation, Jinwei Mao, Wang Luo, Shuang Yang, Daohua Zhu and Zhechen Huang; writing—review and editing, Bao Feng, Jiangtao Xu, Zhechen Huang and Wei Liang; visualization, Jinwei Mao, Bao Feng, Zhechen Huang and Wei Liang; supervision, Wang Luo and Daohua Zhu; project administration, Wang Luo, Jiangtao Xu and Daohua Zhu; funding acquisition, Wang Luo and Daohua Zhu. All authors reviewed and approved the final version of the manuscript.

Availability of Data and Materials: The data and materials are not publicly available.

Ethics Approval: Not applicable.

Conflicts of Interest: The authors declare no conflicts of interest.

References

1. Zhang C, Yang J. Multi-tree genetic programming for adaptive dynamic fault-tolerant task scheduling of satellite edge computing. Future Gener Comput Syst. 2026;175:108099. doi:10.1016/j.future.2025.108099. [Google Scholar] [CrossRef]

2. Ranaweera P, Jurcut AD, Liyanage M. Survey on multi-access edge computing security and privacy. IEEE Commun Surv Tutor. 2021;23(2):1078–124. doi:10.1109/comst.2021.3062546. [Google Scholar] [CrossRef]

3. Tang X, Li X, Yu R, Wu Y, Ye J, Tang F, et al. Digital-twin-assisted task assignment in multi-UAV systems: a deep reinforcement learning approach. IEEE Internet Things J. 2023;10(17):15362–75. doi:10.1109/jiot.2023.3263574. [Google Scholar] [CrossRef]

4. Lou J, Tang Z, Jia W, Zhao W, Li J. Startup-aware dependent task scheduling with bandwidth constraints in edge computing. IEEE Trans Mob Comput. 2023;23(2):1586–600. doi:10.1109/tmc.2023.3238868. [Google Scholar] [CrossRef]

5. Wu G, Chen X, Gao Z, Zhang H, Yu S, Shen S. Privacy-preserving offloading scheme in multi-access mobile edge computing based on MADRL. J Parallel Distrib Comput. 2024;183:104775. doi:10.1016/j.jpdc.2023.104775. [Google Scholar] [CrossRef]

6. Sun F, Zhang Z, Chang X, Zhu K. Toward heterogeneous environment: lyapunov-orientated imphetero reinforcement learning for task offloading. IEEE Trans Netw Serv Manag. 2023;20(2):1572–86. doi:10.1109/tnsm.2023.3266779. [Google Scholar] [CrossRef]

7. Jiang X, Yu FR, Song T, Leung VC. A survey on multi-access edge computing applied to video streaming: some research issues and challenges. IEEE Commun Surv Tutor. 2021;23(2):871–903. doi:10.1109/comst.2021.3065237. [Google Scholar] [CrossRef]

8. Xu X, Liu K, Dai P, Jin F, Ren H, Zhan C, et al. Joint task offloading and resource optimization in NOMA-based vehicular edge computing: a game-theoretic DRL approach. J Syst Archit. 2023;134(1):102780. doi:10.1016/j.sysarc.2022.102780. [Google Scholar] [CrossRef]

9. Zhang C, Yang J. An energy-efficient collaborative offloading scheme with heterogeneous tasks for satellite edge computing. IEEE Trans Netw Sci Eng. 2024;11(6):6396–407. doi:10.1109/tnse.2024.3476968. [Google Scholar] [CrossRef]

10. Huang X, Huang G. Joint optimization of energy and task scheduling in wireless-powered irs-assisted mobile-edge computing systems. IEEE Internet Things J. 2023;10(12):10997–13. doi:10.1109/jiot.2023.3242951. [Google Scholar] [CrossRef]

11. Deng Y, Chen X, Zhu G, Fang Y, Chen Z, Deng X. Actions at the edge: jointly optimizing the resources in multi-access edge computing. IEEE Wirel Commun. 2022;29(2):192–8. doi:10.1109/mwc.006.2100699. [Google Scholar] [CrossRef]

12. Porambage P, Okwuibe J, Liyanage M, Ylianttila M, Taleb T. Survey on multi-access edge computing for internet of things realization. IEEE Commun Surv Tutor. 2018;20(4):2961–91. doi:10.1109/comst.2018.2849509. [Google Scholar] [CrossRef]

13. Liu J, Yang P, Chen C. Intelligent energy-efficient scheduling with ant colony techniques for heterogeneous edge computing. J Parall Distrib Comput. 2023;172:84–96. doi:10.1016/j.jpdc.2022.10.003. [Google Scholar] [CrossRef]

14. Deng X, Yin J, Guan P, Xiong NN, Zhang L, Mumtaz S. Intelligent delay-aware partial computing task offloading for multiuser industrial Internet of Things through edge computing. IEEE Internet Things J. 2021;10(4):2954–66. doi:10.1109/jiot.2021.3123406. [Google Scholar] [CrossRef]

15. Li Y, Liu B, Wu E, Li J, Zhou Z, Zhang W. DRA-MQoS: an MQoS scheduling algorithm based on resource feature matching in federated edge cloud. Concurr Comput Pract Exp. 2023;35(2):e7478. doi:10.1002/cpe.7478. [Google Scholar] [CrossRef]

16. Liu Q, Zeng L, Bilal M, Song H, Liu X, Zhang Y, et al. A multi-swarm PSO approach to large-scale task scheduling in a sustainable supply chain datacenter. IEEE Trans Green Commun Netw. 2023;7(4):1667–77. doi:10.21203/rs.3.rs-1844078/v1. [Google Scholar] [CrossRef]

17. Gao A, Zhang S, Zhang Q, Hu Y, Liu S, Liang W, et al. Task offloading and energy optimization in hybrid UAV-assisted mobile edge computing systems. IEEE Trans Veh Technol. 2024;73(8):12052–66. doi:10.1109/tvt.2024.3380003. [Google Scholar] [CrossRef]

18. Cai J, Liu W, Huang Z, Yu FR. Task decomposition and hierarchical scheduling for collaborative cloud-edge-end computing. IEEE Trans Serv Comput. 2024;17(6):4368–82. doi:10.1109/tsc.2024.3402169. [Google Scholar] [CrossRef]

19. Chen W, Wei X, Chi K, Yu K, Tolba A, Mumtaz S, et al. Computation time minimized offloading in NOMA-enabled wireless powered mobile edge computing. IEEE Trans Commun. 2024;72(11):7182–97. doi:10.1109/tcomm.2024.3405316. [Google Scholar] [CrossRef]

20. Fan W, Zhang Y, Zhou G, Liu Y. Deep reinforcement learning-based task offloading for vehicular edge computing with flexible RSU-RSU cooperation. IEEE Trans Intell Transp Syst. 2024;25(7):7712–25. doi:10.1109/tits.2024.3349546. [Google Scholar] [CrossRef]

21. Guo H, Chen X, Zhou X, Liu J. Trusted and efficient task offloading in vehicular edge computing networks. IEEE Trans Cogn Commun Netw. 2024;10(6):2370–82. doi:10.1109/globecom48099.2022.10000816. [Google Scholar] [CrossRef]

22. Li K, Wang X, He Q, Wang J, Li J, Zhan S, et al. Computation offloading in resource-constrained multi-access edge computing. IEEE Trans Mob Comput. 2024;23(11):10665–77. doi:10.1109/tmc.2024.3383041. [Google Scholar] [CrossRef]

23. Li L, Qiu Q, Xiao Z, Lin Q, Gu J, Ming Z. A two-stage hybrid multi-objective optimization evolutionary algorithm for computing offloading in sustainable edge computing. IEEE Trans Consum Electron. 2024;70(1):735–46. doi:10.1109/tce.2024.3376930. [Google Scholar] [CrossRef]

24. Lin H, Yang L, Guo H, Cao J. Decentralized task offloading in edge computing: an offline-to-online reinforcement learning approach. IEEE Trans Comput. 2024;73(6):1603–15. doi:10.1109/tc.2024.3377912. [Google Scholar] [CrossRef]

25. Liu Q, Zhang H, Zhang X, Yuan D. Improved DDPG based two-timescale multi-dimensional resource allocation for multi-access edge computing networks. IEEE Trans Veh Technol. 2024;73(6):9153–8. doi:10.1109/tvt.2024.3360943. [Google Scholar] [CrossRef]

26. Liu P, An K, Lei J, Sun Y, Liu W, Chatzinotas S. Computation rate maximization for SCMA-aided edge computing in IoT networks: a multi-agent reinforcement learning approach. IEEE Trans Wirel Commun. 2024;23(8):10414–29. doi:10.1109/twc.2024.3371791. [Google Scholar] [CrossRef]

27. Neto JLD, Yu SY, Macedo DF, Nogueira JMS, Langar R, Secci S. ULOOF: a user level online offloading framework for mobile edge computing. IEEE Trans Mob Comput. 2018;17(11):2660–74. doi:10.1109/tmc.2018.2815015. [Google Scholar] [CrossRef]

28. Han C, Yao H, Mai T, Zhang N, Guizani M. QMIX aided routing in social-based delay-tolerant networks. IEEE Trans Veh Technol. 2021;71(2):1952–63. doi:10.1109/tvt.2021.3133449. [Google Scholar] [CrossRef]

29. Peng B, Rashid T, Schroeder de Witt C, Kamienny PA, Torr P, Boehmer W, et al. FACMAC: factored multi-agent centralised policy gradients. Adv Neural Inf Process Syst. 2021;34:12208–21. [Google Scholar]

30. Yu C, Velu A, Vinitsky E, Gao J, Wang Y, Bayen A, et al. The surprising effectiveness of PPO in cooperative multi-agent games. Adv Neural Inf Process Syst. 2022;35:24611–24. [Google Scholar]

31. Lin H, Xiao B, Zhou X, Zhang Y, Liu X. A multi-tier offloading optimization strategy for consumer electronics in vehicular edge computing. IEEE Trans Consum Electron. 2025;71(1):2118–30. doi:10.1109/tce.2025.3527043. [Google Scholar] [CrossRef]


Cite This Article

APA Style
Mao, J., Luo, W., Xu, J., Zhu, D., Liang, W. et al. (2026). Heterogeneous Computing Power Scheduling Method Based on Distributed Deep Reinforcement Learning in Cloud-Edge-End Environments. Computers, Materials & Continua, 87(2), 85. https://doi.org/10.32604/cmc.2026.072505
Vancouver Style
Mao J, Luo W, Xu J, Zhu D, Liang W, Huang Z, et al. Heterogeneous Computing Power Scheduling Method Based on Distributed Deep Reinforcement Learning in Cloud-Edge-End Environments. Comput Mater Contin. 2026;87(2):85. https://doi.org/10.32604/cmc.2026.072505
IEEE Style
J. Mao et al., “Heterogeneous Computing Power Scheduling Method Based on Distributed Deep Reinforcement Learning in Cloud-Edge-End Environments,” Comput. Mater. Contin., vol. 87, no. 2, pp. 85, 2026. https://doi.org/10.32604/cmc.2026.072505


cc Copyright © 2026 The Author(s). Published by Tech Science Press.
This work is licensed under a Creative Commons Attribution 4.0 International License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
  • 306

    View

  • 117

    Download

  • 0

    Like

Share Link