Abstract
Quantum computing and machine learning convergence enable powerful new approaches for optimizing mobile edge computing (MEC) networks. This paper uses Lyapunov optimization theory to propose a novel quantum machine learning framework for stabilizing computation offloading in next-generation MEC systems. Our approach leverages hybrid quantum-classical neural networks to learn optimal offloading policies that maximize network performance while ensuring the stability of data queues, even under dynamic and unpredictable network conditions. Rigorous mathematical analysis proves that our quantum machine learning controller achieves close-to-optimal performance while bounding queue backlogs. Extensive simulations demonstrate that the proposed framework significantly outperforms conventional offloading approaches, improving network throughput by up to 30% and reducing power consumption by over 20%. These results highlight the immense potential of quantum machine learning to revolutionize next-generation MEC networks and support emerging applications at the intelligent network edge.
Similar content being viewed by others
Introduction
The MEC is an innovative concept emerging due to the extension of the advantages of conventional cloud computing to near-ubiquitous wireless networks1. The proposed concept known as MEC can bring a rethink in the utilization of the computation capability of the mobile terminal and extend new computationally intensive applications, including AR, intelligent IoT, and autonomous systems2,3. Nevertheless, computation offloading in MEC networks faces various challenges due to the wireless edge environment’s characteristics and randomness. Computation offloading needs to manage resources such as radio spectrum, computing capability, and energy to achieve a desired tradeoff between latency, throughput, and power4. This is impossible in MEC networks due to the active Nature of the wireless channel condition and the computation load5. Aggressive offloading can result in network overload and computation demands affecting user experience. Thus, offloading too conservatively leads to not utilizing more edge resources to the maximum. Thus, it is clear that to realize a good and stable MEC network, it is necessary to have a flexible offloading strategy suitable for dynamic networks.
Stochastic optimization is an effective paradigm from stochastic network optimization theory that would allow the formulation of online control protocols with guarantees in performance6. It has been successfully used for queuing system stability in different networking environments7,8. The overall strategy is to make control verdicts that minimize an expression of the form drift plus penalty as greedily as possible, steering the system queues to a desired operational point for a selected goal. Nevertheless, advanced techniques founded on Lyapunov optimization remain challenging in high-dimensional MEC network environments with state/action space and nonlinear dynamics. Later, a completely new paradigm of computing based on the principles of quantum mechanics called quantum computing came to the front, and the quantum model of computing will offer exponential solutions for some of the computational problems9. The studies conducted in the last years indicate that due to the implementation of quantum features such as the principle of superposition, quantum computers can solve optimization problems challenging for classical systems10. On the same note, there has been significant improvement in machine learning, particularly about learning from data and learning of arbitrary functions. We see exciting opportunities for enhancing wireless networks and offerings emerging from the recently evolved cross-discipline of quantum machine learning (QML).
A quantum convolutional neural network (QCNN) algorithm uses minimal parameters for quantum state processing, effectively recognizing topological phases and improving error correction in quantum systems11. Introduces a quantum recurrent neural network (qRNN) using arrays of Rydberg atoms for quantum computing. It demonstrates how qRNNs can perform cognitive tasks like multitasking, decision-making, and memory storage by leveraging quantum dynamics. The system shows advantages over classical RNNs through quantum interference, measurement basis flexibility, and efficient stochastic process simulation12. QCapsNet, a quantum version of capsule networks, uses quantum dynamic routing for enhanced classification accuracy and potential explainability in quantum machine learning, outperforming conventional quantum classifiers13. A quantum neuroevolution algorithm uses evolutionary principles to optimize quantum neural networks through graph-based path searching, demonstrating effectiveness in image classification and topological state recognition tasks14.
A quantum protocol enables secure distributed learning by combining blind quantum computation with differential privacy, allowing multiple parties to train shared networks while protecting private data15. Quantum AI combines AI tools for quantum problems and quantum computing for enhanced AI capabilities. However, classical and quantum learning systems are vulnerable to adversarial attacks through carefully crafted input perturbations that cause incorrect predictions16. High expressibility in quantum tangent kernel models can lead to exponential concentration to zero, affecting global and local loss functions. This concentration issue persists with local encoding schemes, impacting wide quantum variational circuit design17. A quantum classifier demonstrates continual learning across multiple tasks while avoiding catastrophic forgetting through elastic weight consolidation, achieving 92.3% accuracy18. Quantum Machine Learning (QML) is particularly well-suited for the MEC offloading problem because it can efficiently handle the high-dimensional state and action spaces inherent in complex network optimization. The quantum neural network architecture leverages quantum superposition and parallelism to evaluate multiple offloading decisions simultaneously, providing a quadratic speedup over classical approaches for large-scale networks. While the proposed framework can be simulated on classical computers using hybrid quantum-classical simulators for the current implementation, it is designed to take full advantage of quantum hardware as it becomes available. The QML approach demonstrates superior performance with up to 30% higher throughput and 20% lower energy consumption compared to classical methods, particularly in handling dynamic network conditions. The quantum-enhanced training algorithm also achieves significant speedups in computing gradient moments, reducing training time complexity from O(n2) to O(n) for n-dimensional inputs.
Motivation and background.
Mobile Edge Computing (MEC) has emerged as a transformative paradigm that extends cloud computing capabilities to the wireless network edge. This architectural shift supports computation-intensive applications like augmented reality, intelligent IoT, and autonomous systems by bringing processing power closer to end users.
Challenges.
Several critical challenges need to be addressed for effective MEC deployment:
Resource Management Complexity The wireless edge environment introduces significant randomness and variability in channel conditions, computation loads, and resource availability. Managing radio spectrum, computing capability, and energy resources simultaneously becomes extremely challenging.
Network Stability Issues Aggressive offloading can lead to congestion and degraded user experience, while conservative offloading results in the underutilization of edge resources. Finding the right balance requires sophisticated control mechanisms.
Dynamic Network Conditions The highly dynamic Nature of wireless channels and computation demands makes it difficult to maintain consistent performance. Traditional static optimization approaches need help to adapt to these changing conditions.
This research work proposes a novel QML framework to stabilize computation offloading and resource allocation in next-generation MEC networks based on Lyapunov optimization theory. The key innovations are:
-
1.
We develop hybrid quantum-classical neural network models to learn effective offloading policies by mapping the system state to offloading actions. The quantum layers leverage efficient quantum linear algebra subroutines to speed training and inference.
-
2.
We mathematically incorporate Lyapunov drift-plus-penalty expressions into the loss function of the QML models. This allows the models to learn offloading policies that jointly stabilize data queues while optimizing throughput and energy efficiency.
-
3.
We provide a rigorous theoretical analysis using Lyapunov optimization theory to prove the proposed QML framework’s stability and near-optimal performance. The analysis provides bounded guarantees on queue backlogs and throughput utility.
-
4.
Extensive simulations demonstrate the significant performance gains of our approach over both conventional optimization algorithms and pure classical deep learning baselines. The QML framework improves network throughput by up to 30% while reducing energy consumption by over 20%.
This is the first work to present a QML approach for stabilizing MEC networks using the Lyapunov optimization theory. Quantum computing and machine learning are used to build a solid and powerful framework that can quickly enable next-generation MEC systems. The outcomes can affect a variety of new application domains based on smart edge computing and networking.
Our research paper’s structure is as follows: “Related work” presents related work, which includes MEC, Lyapunov optimization, and quantum machine learning. Section “System model and problem formulation” gives the system model and problem formulation. Section “Quantum machine learning framework” presents the proximal message passing-based QML framework and training algorithm based on Lyapunov optimization. We outline the theoretical analysis of the performance bounds in “Theoretical analysis”. Section “Simulation results and discussion” gives the simulation results and the discussion on the same. Section “Quantum circuit implementation details” provides the Quantum Circuit Implementation Details, “Comparative performance analysis” Comparative Performance Analysis, and “Conclusion” closes the paper.
Related work
Mobile edge computing
MEC has attracted significant research interest in recent years as a key enabling technology for 5G and beyond1,2. By moving computing resources to the network edge, MEC can provide low-latency, high-bandwidth access to powerful computation capabilities for resource-constrained mobile devices3. This is crucial for supporting emerging applications such as augmented reality, intelligent IoT, autonomous systems, and tactile Internet4,5.
Computation offloading is one of the primary approaches in MEC systems, where a mobile device transfers the Execution of a particular task to a distant server at the edge6. It can boost the available computing capability for portable devices and extend battery lifespan7. However, computation offloading comes accompanied by new issues, such as joint communication and computation resource management under dynamic network conditions8. To address these issues, diverse optimization techniques have been developed inclu, ding convex optimization9, game theory10, and reinforcement learning. Proposes an intelligent game anti-interference computing model for metaverse security, using a Stackelberg game framework where source devices lead, collaborative devices sub-lead, and jammers follow. The SGPL algorithm is designed to combat jamming in 5G networks through differential dynamics and policy-based learning19.
However, most existing works focus on optimizing computation offloading under a static network environment with known system statistics. In contrast, practical MEC systems exhibit highly dynamic and unpredictable channel conditions, workload arrivals, and resource availability variations. This requires adaptive online control policies to learn and respond to the changing environment. Our work addresses this challenge by leveraging QML to learn stable and optimal offloading policies. Table 1 gives the comparison of critical aspects of the related works:
Lyapunov optimization
Lyapunov optimization is a robust stochastic network optimization theory tool for designing online control algorithms with provable performance guarantees6. It has been widely applied in communication networks, power systems, and data center networks. The key idea is to define a Lyapunov function, typically a weighted sum of squares of queue backlogs, that represents the aggregate congestion in the system. The control actions are then chosen to greedily minimize a drift-plus-penalty expression, which pushes the Lyapunov function towards a small value while optimizing a performance objective.
Lyapunov optimization provides a solid theoretical framework for trading off utility optimality and queue stability. It has been used to stabilize queues in various MEC offloading scenarios. However, classical Lyapunov-based approaches struggle in complex network settings with high-dimensional state/action spaces and highly nonlinear dynamics. They also rely on the instantaneous optimization of a drift-plus-penalty expression, which can be computationally expensive. Our work addresses these limitations by integrating Lyapunov optimization with QML to learn optimized control policies offline.
Quantum machine learning
Quantum Machine Learning is a relatively new field of study that uses quantum computing to create new methods for machine learning. The beauty of deploying on quantum states capable of describing exponentially large data sets allows QML to reduce the time taken to train and infer the complex model. There are several kinds of centroid QML models, such as quantum neural networks, quantum kernel methods, and quantum Boltzmann machines34,35,36.
Recently, there has been growing interest in applying QML to enhance wireless networks and communications. For example, quantum convolutional neural networks have been used for modulation classification and signal detection in wireless systems. Quantum reinforcement learning has been applied to solve network resource allocation problems. This is the first work to leverage QML for stabilizing computation offloading in MEC based on Lyapunov optimization theory. Our approach harnesses the power of quantum neural networks to learn complex offloading policies that optimize MEC performance34,37.
System model and problem formulation
MEC system model
Consider a MEC network consisting of a set of M mobile devices and a set \(\:\mathcal{o}\mathcal{f}\mathcal{\:}\mathcal{S}\)of S edge servers, as shown in Fig. 1. Each mobile device \(\:m\in\:\mathcal{M}\)generates computation tasks over time that can be processed locally or offloaded to an edge server. The edge servers have significantly higher computation capacities than mobile devices30.
The model operates in discrete time slots \(\:t\in\:\text{0,1},2,…\:\) In each time slot t, mobile device m generates computation tasks according to self-regulating and identically distributed. Process \(\:{A}_{m}\left(t\right)\) with arrival rate \(\:{\lambda\:}_{m}\). The no. of CPU cycles characterizes each task used for Execution. Let \(\:{D}_{m}\left(t\right)\)denote the number of CPU cycles arriving at device m in slot t. The tasks are queued in separate buffers at each mobile device, with \(\:{Q}_{m}\left(t\right)\:\)denoting the queue backlog of device m in slot t.
The wireless channel conditions between mobile devices and edge servers vary randomly over time due to factors like device mobility and environmental changes. Let \(\:{h}_{ms}\left(t\right)\:\)denote the channel state between mobile device m and edge server s in slot t, determining the transmission data rate.
Mobile devices make offloading decisions in each time slot to determine how many task bits to offload to each edge server. Let \(\:{\mu\:}_{ms}\left(t\right)\:\)denote the number of task bits offloaded from mobile device m to server s in slot t. The edge servers allocate their computation resources to execute the offloaded tasks. Let\(\:{}_{}\left(\right)\)’s note the CPU frequency servers allocate in slot t to process the offloaded workload.
Computation model
Computation tasks can be processed locally on mobile devices or remotely on edge servers. For local Execution, the processing latency is determined by the mobile device’s CPU frequency39. Let \(\:{f}_{m}^{loc}\:\)denote the local computation capability of device m in CPU cycles per second. Then, the time required to process q task bits locally at device m is given by:
where L denotes the number of CPU cycles required per bit.
For edge execution, the processing delay can be separated into the transmission and computation delays. The transmission delay is determined by the wireless channel characteristics and the amount of data being offloaded. W represents the system’s channel bandwidth, whereas Pm denotes the mobile device’s transmission power. The Shannon capacity determines the transmission data rate between mobile device m and server s in slot t40:
where \(\:{N}_{0}\:\) is the noise power spectral density.
The computation delay at the edge server depends on the allocated CPU frequency \(\:{f}_{s}\left(t\right)\:\). Similar to local Execution, the processing time for q task bits at server s is:
The total delay for edge execution also includes the round-trip transmission delay:
Problem formulation
The overall goal is to design an online computation offloading policy that jointly optimizes the system performance while maintaining the stability of the task queues. The offloading policy \pi consists of the offloading decisions \(\:{\mu\:}_{ms}\left(t\right)\:\)made by each mobile device m to each edge server s in each time slot t.
We define the following optimization objectives:
-
1.
Throughput Maximization: Maximize the long-term average number of task bits processed in the system, given by41:
$$\:\stackrel{-}{R}\left(\pi\:\right)=\underset{T\to\:{\infty\:}}{lim}\frac{1}{T}\sum\:_{t=0}^{T-1}\mathbb{E}\left[\sum\:_{m=1}^{M}\sum\:_{s=1}^{s}s{\mu\:}_{ms}\left(t\right)\right]$$(5) -
2.
Power minimization: minimize the long-term average power consumption of the mobile devices, given by:
$$\:\stackrel{-}{P}\left(\pi\:\right)=\underset{T\to\:{\infty\:}}{lim}\frac{1}{T}\sum\:{t=0}^{T-1}\mathbb{E}\left[\sum\:_{m=1}^{M}\sum\:_{s=1}^{S}\frac{\mu\:ms\left(t\right)}{rms\left(t\right)}{P}_{m}\right]$$(6)
The expectations are regarding the randomness of the task arrivals, channel conditions, and the offloading policy.
We formulate a utility maximization problem that combines the above objectives:
where V is a positive weight that determines the relative importance of throughput and power, and \(\:{\stackrel{-}{Q}}_{m}\:\) is the time-averaged expected queue backlog:
The constraint \(\:{\stackrel{-}{Q}}_{m}<{\infty\:}\) ensures the stability of the task queues, i.e., the task buffers should not grow unboundedly over time.
Problem P1 is challenging to solve due to several reasons. First, the objective function couples the offloading decisions across multiple time slots through the time-averaged expectations. Second, the system dynamics are stochastic and unpredictable due to the randomness in the task arrivals and channel conditions. Third, the stability constraint requires the offloading policy to adapt to the instantaneous system state while ensuring long-term queue stability. Finally, the optimization variables \(\:{\mu\:}_{ms}\left(t\right)\:\)are integer-valued, making the problem non-convex.
In the next section, we develop a novel quantum machine learning framework based on Lyapunov optimization theory to tackle this problem. The key idea is to leverage the power of quantum neural networks to learn effective offloading policies that satisfy the Lyapunov stability constraints.
Quantum machine learning framework
Lyapunov optimization
To solve problem P1, we first introduce a set of virtual queues \(\:Hm\left(t\right)\) for each mobile device m, which evolve as follows:
where \(\:{\gamma\:}_{m}\:\)is a perturbation parameter that determines the desired throughput level for device m, intuitively, the virtual queue Hm(t) accumulates the difference between the actual throughput\(\:\sum\:_{s=1}^{S}{\mu\:}_{ms}\left(t\right)\) and the desired throughput \(\:{\gamma\:}_{m}\:\)over time. Stabilizing \(\:Hm\left(t\right)\) ensures that the actual throughput approaches the desired level.
We define the Lyapunov function as:
The Lyapunov function represents the aggregate congestion in the system, considering both the actual task queues and the virtual queues. The one-step Lyapunov drift is given by:
\(\:\mathbf{Q}\left(t\right)\) and \(\:\mathbf{H}\left(t\right)\) are vectors of all actual and virtual queue backlogs, respectively.
Following the Lyapunov optimization approach, we define the drift-plus-penalty expression:
where \(\:R\left(t\right)=\sum\:_{m=1}^{M}\:\sum\:_{s=1}^{S}\:\mu\:ms\left(t\right)\) is the instantaneous throughput and
is the instantaneous power consumption.
The key idea of Lyapunov optimization is to make control decisions that minimize the drift-plus-penalty expression in each time slot. This approach simultaneously pushes the system queues towards stability while optimizing the objective function.
After some mathematical manipulations (detailed derivation omitted for brevity), we can obtain an upper bound on the drift-plus-penalty expression:
where B is a constant, and \(\:{{\Psi\:}}_{ms}\left(t\right)\:\)is defined as:
The term \(\:{\Psi\:}ms\left(t\right)\)can be interpreted as a weight determining the priority of offloading tasks from device m to server s in slot t. A larger \(\:{\Psi\:}ms\left(t\right)\:\)indicates a higher priority for offloading.
Quantum neural network architecture
To minimize the upper bound on the drift-plus-penalty expression, we propose a quantum neural network (QNN) architecture that learns to map the system state to optimal offloading decisions. The QNN takes as input the current queue backlogs \(\:\mathbf{Q}\left(t\right),\:\)virtual queue backlogs \(\:\mathbf{H}\left(t\right),\:\)and channel states \(\:\mathbf{h}\left(t\right),\:\)and outputs the offloading decisions \(\:{\mu\:}_{ms}\left(t\right)\:\)for all device-server pairs.
The proposed QNN architecture consists of hybrid quantum-classical layers, as shown in Fig. 2. The input layer is classical, followed by several quantum convolutional layers, quantum pooling layers, and finally classical fully connected layers. The QNN’s decision-making process maps system states (queue backlogs, channel conditions) to offloading actions through quantum convolutional layers that extract relevant features. The network learns to prioritize devices with larger queue backlogs and better channel conditions, while the Lyapunov optimization ensures stability by minimizing the drift-plus-penalty expression. The quantum layers enable efficient exploration of high-dimensional state spaces to find optimal offloading policies.
The quantum convolutional layers leverage quantum circuits to perform convolution operations on the input data. Each quantum convolutional layer consists of a set of parameterized quantum circuits that act as filters. The quantum circuits are composed of single-qubit rotation gates and two-qubit entangling gates, as shown in Fig. 3.
The quantum pooling layers perform dimensionality reduction using quantum measurements. We use a quantum average pooling operation that computes the expectation value of a Pauli-Z observable on groups of qubits.
The final classical layers map the quantum measurement outcomes to the offloading decisions \(\:{\mu\:}_{ms}\left(t\right).\:\)We use rectified linear unit (ReLU) activation functions in the hidden layers and a sigmoid activation in the output layer to ensure non-negative offloading decisions.
Figure 4 depicts the overall quantum circuit structure for the QNN layers, including the input encoding, quantum convolutional layers, quantum pooling layers, and classical output layers. The hybrid quantum-classical architecture allows leveraging quantum speedups while integrating with classical processing.
Table 2 specifies the quantum neural network (QNN) architecture:
-
The input features are pre-processed and encoded into quantum states using amplitude encoding.
-
Each quantum convolutional layer applies the quantum filters in parallel to the input quantum state.
-
Quantum pooling layers compute expectation values of Pauli-Z operators to reduce dimensionality.
-
Classical layers use standard dense connections with weight matrices and bias vectors.
-
The output layer generates the offloading decision values µ(t) for each device-server pair.
This hybrid quantum-classical architecture allows leveraging quantum speedups while integrating with classical neural network layers. The quantum convolutional layers efficiently extract relevant features from the input quantum states, while the classical layers provide additional processing and help generate the final offloading decisions.
Quantum-enhanced training algorithm
To train the QNN, we propose a quantum-enhanced version of the famous Adam optimization algorithm. The key idea is to leverage quantum linear algebra subroutines to speed up the computation of gradients and parameter updates.
The loss function for training the QNN is designed to minimize the upper bound on the drift-plus-penalty expression:
The gradients of the loss function with respect to the QNN parameters are computed using the parameter shift rule, which is compatible with quantum backpropagation.
The quantum-enhanced Adam algorithm uses quantum linear algebra subroutines to compute the gradients’ first and second moment estimates efficiently. Specifically, we leverage the quantum singular value estimation (QSVE) algorithm to speed up matrix operations involved in the moment updates.
Figure 5 outlines the proposed quantum-enhanced training procedure.
The Quantum Matrix Multiply function uses the QSVE algorithm to perform matrix multiplication efficiently in the quantum domain. This provides a quadratic speedup over classical matrix multiplication for large matrices.
% Quantum-Enhanced MEC Offloading Algorithm.
% Initialize parameters numDevices = M; % Number of mobile devices numServers = S; % Number of edge servers V = 1000; % Lyapunov parameter maxEpochs = 1000;
% Initialize QNN parameters theta = initializeQuantumParameters(); qnn = struct(‘theta’, theta);
% Initialize queues Q = zeros(numDevices, 1); % Task queues H = zeros(numDevices, 1); % Virtual queues.
% Training loop function trainQNN() for epoch = 1:maxEpochs % Get current system state state = getSystemState();

end.
% Online offloading algorithm function onlineOffloading() while true % Observe current system state queues = getQueueBacklogs(); channels = getChannelStates();

end.
% Helper functions function loss = computeLyapunovLoss(decisions, Q, H) drift = sum(Q.^2 + H.^2); penalty = V * computeUtility(decisions); loss = drift - penalty; end.
function state = encodeQuantumState(classicalState) % Encode classical state into quantum state using amplitude encoding normState = normalizeState(classicalState); state = quantumAmplitudeEncoding(normState); end.
Figure 6 show the Lyapunov drift-plus-penalty training loss over 1000 epochs, and the bottom heatmap displays the final offloading decisions for 10 mobile devices across 3 edge servers.
Table 3 shows hyperparameters used for the Quantum Machine Learning (QML) training.
This table provides a comprehensive overview of all hyperparameters used in training the QML model for computation offloading in the MEC network. The parameters were chosen to balance computational efficiency with model performance while ensuring stable convergence of the quantum-classical hybrid architecture.
Figure 7 presents the convergence curve of the QML training process, plotting the Lyapunov drift-plus-penalty loss over training epochs. It demonstrates the stable and efficient convergence of the proposed quantum-enhanced training algorithm.
Online offloading algorithm
Once the QNN is trained, we use it to make online offloading decisions in each time slot. Algorithm 2 outlines the online offloading procedure.
Figure 8 Flowchart of the online offloading algorithm showing the step-by-step decision process. In each time slot, the algorithm observes the current system state, computes the offloading weights Ψ(t), and uses the trained QNN to determine the offloading decisions µ(t). The offloading decisions are then executed, and the queues are updated accordingly.
System state observation.
The current system state vector S(t) at time t:
where Q(t): Task queue backlog vector; H(t): Virtual queue backlog vector; α(t): Channel state vector.
Offloading weight computation.
The offloading weight Ψms(t) for device m to server s:
-
Rms(t): Achievable transmission rate.
-
Ems(t): Energy consumption.
-
V: Lyapunov control parameter.
QNN input processing.
The normalized input vector X(t):
Offloading decision output.
The offloading decision vector µ(t):
θ: Trained QNN parameters.
Queue updates.
Task queue update:
Virtual queue update:
where [x]⁺: max(x,0); \(\:{A}_{m}\left(t\right)\:\): Task arrivals; \(\:{\gamma\:}_{m}\:\): Target throughput threshold; \(\:{\mu\:}_{ms}\left(t\right)\) ∈ [0,1]: Offloading decision for device m to server s.
These equations mathematically describe each step in the flowchart for making online offloading decisions in the MEC network using the trained quantum neural network.
% Online Offloading Algorithm for MEC Networks function onlineOffloading() % Initialize parameters M = 10; % Number of mobile devices S = 3; % Number of edge servers V = 1000; % Lyapunov parameter.

end.
% Helper Functions function R = getTransmissionRate(m, s, alpha) % Calculate transmission rate using Shannon capacity W = 20e6; % Bandwidth P = 0.1; % Transmission power N0 = 1e-9; % Noise power R = W log2(1 + Palpha(m, s)/N0); end.
function E = getEnergyConsumption(m, s) % Calculate energy consumption for offloading P_tx = 0.1; % Transmission power t_tx = 0.001; % Transmission time E = P_tx * t_tx; end.
function X = normalizeState(state) % Normalize system state for QNN input X = (state - mean(state)) ./ std(state); end.
function executeOffloadingDecisions(mu) % Execute computed offloading decisions for m = 1:size(mu,1) for s = 1:size(mu,2) if mu(m, s) > 0 offloadTasksToServer(m, s, mu(m, s)); end end end end.
Theoretical analysis
In this section, we provide theoretical analysis to prove the proposed QML framework’s stability and performance guarantees. We first establish the queue stability and then derive bounds on the achievable utility.
Queue stability
We begin by proving the stability of both the actual and virtual task queues.
Theorem 1
(Queue Stability) Under the proposed QML offloading policy, the actual task queues and virtual queues are mean-rate stable, i.e.,
Proof
The proof follows from the Lyapunov drift analysis. Let Δ(t) be the one-step Lyapunov drift as defined earlier. We can show that:
Dividing both sides by T and taking the limit as T → ∞, we get:
Since L(T) is a sum of squared queue backlogs, this implies that both \(\:{Q}_{m}\left(T\right)\:\)and \(\:{H}_{m}\left(T\right)\:\)are mean-rate stable.
Performance bounds
Next, we derive bounds on the achievable utility of the proposed QML framework.
Theorem 2
(Utility Bound) The proposed QML offloading policy achieves a utility that is within O(1/V) of the optimal utility:
R and P are the optimal throughput and power consumption achievable by any policy.
Proof
From the Lyapunov optimization theory, we can show that the drift-plus-penalty expression satisfies:
Summing over T time slots and dividing by VT, we get:
Taking the limit as T → ∞ and using the fact that L(T)/T → 0 (from Theorem 1), we obtain the desired result.
Theorem 2 shows that by increasing the parameter V, we can push the utility arbitrarily close to the optimal value, at the cost of increased queue backlogs (which grow as O(V) according to Theorem 1). This demonstrates the [O(1/V), O(V)] utility-backlog tradeoff that is characteristic of Lyapunov optimization approaches.
Quantum speedup
Finally, we analyze the quantum speedup achieved by our QML framework compared to classical approaches.
Theorem 3
(Quantum Speedup) The proposed quantum-enhanced training algorithm achieves a quadratic speedup in the computation of gradient moments compared to classical training, reducing the time complexity from O(n2) to O(n) for input dimension n.
Proof
The key quantum speedup comes from using the quantum singular value estimation (QSVE) algorithm in the moment updates of the Adam optimizer. Classical matrix multiplication for moment updates has time complexity O(n2) for n-dimensional inputs. The QSVE algorithm can perform the same operation with complexity O(n), providing a quadratic speedup.
This quantum speedup enables faster neural network training, especially for high-dimensional system states in large-scale MEC networks. The faster training allows the offloading policy to adapt more quickly to changing network conditions.
Simulation results and discussion
We conducted extensive simulations to evaluate the performance of the proposed QML framework for computation offloading in MEC networks. We compared our approach against the following baselines:
-
1.
Local Execution: All tasks are processed locally at the mobile devices.
-
2.
Random Offloading: Tasks are randomly offloaded to edge servers with equal probability.
-
3.
Greedy Offloading: Tasks are always offloaded to the server with the best channel condition.
-
4.
Classical Deep Learning: A classical deep neural network trained to make offloading decisions.
-
5.
Lyapunov Optimization: A conventional Lyapunov-based algorithm that solves an optimization problem in each time slot.
Simulation setup
We created a MEC network using ten mobile devices and three edge servers. The task arrival process at each mobile device follows a Poisson distribution, with an average rate of λ = 5 tasks per slot. The wireless channel gains are calculated using a Rayleigh fading model. The simulation parameters are given in Table 4.
The QNN architecture consists of 2 quantum convolutional layers, 2 quantum pooling layers, and 2 classical fully connected layers. We used 4 qubits for each quantum filter in the convolutional layers. The QNN was trained for 1000 epochs using the quantum-enhanced Adam optimizer with a learning rate of 0.001.
Table 5 shows the computational complexity between quantum and classical algorithms for the MEC offloading optimization problem:
Advantages of quantum algorithms:
-
Exponential speedup for policy optimization through quantum superposition.
-
Quadratic speedup for resource allocation via quantum parallelism.
-
Significant reduction in training time through quantum gradient descent.
-
More efficient exploration of high-dimensional state/action spaces.
This comparison demonstrates the theoretical advantages of the quantum-enhanced approach over classical methods, particularly for large-scale MEC networks with many devices and servers.
Performance evaluation
The proposed QML framework significantly outperformed all baseline methods in our simulations. As shown in Fig. 9, the QML approach achieved up to 30% higher throughput compared to the next best algorithm (Lyapunov Optimization). Figure 10 demonstrates the QML framework’s superior performance across key metrics including throughput, power consumption, and latency. The intelligent offloading decisions learned by the QML model enabled better utilization of edge resources under dynamic network conditions. Table 5 summarizes the comprehensive improvements, with the QML approach reducing average latency by 25%, energy consumption by 22%, and queue backlogs by 18% compared to conventional techniques. These results highlight the effectiveness of our quantum-enhanced approach for next-generation MEC networks.
Table 6 demonstrates the comprehensive improvements achieved by the QML framework across multiple key performance metrics, highlighting its effectiveness for next-generation MEC networks.
Figure 9 compares the performance of different offloading methods, including the proposed QML approach, in terms of throughput, power consumption, and latency. It demonstrates the superior performance of the QML framework compared to baseline methods.
Figure 10 compares key performance metrics across different offloading methods, such as throughput, power consumption, and latency. It highlights the significant improvements achieved by the QML framework over conventional approaches.
Figure 11 heatmap visualizes the offloading decisions made by the QML framework across multiple time slots for different mobile devices and edge servers. It provides insights into the dynamic adaptation of the offloading policy to varying network conditions.
Figure 12 plots the average system throughput achieved by different offloading schemes as a function of the number of time slots. It demonstrates the significant throughput improvements of the QML approach compared to baseline methods, achieving up to 30% higher throughput.
Figure 13 compares the average power consumption of mobile devices under different offloading schemes. The QML approach achieves the lowest power consumption, reducing energy usage by over 20% compared to classical deep learning and conventional Lyapunov optimization. This demonstrates the effectiveness of the QML framework in jointly optimizing throughput and energy efficiency.
To evaluate the task queues’ stability, we plot the average queue backlog over time in Fig. 14. The QML approach maintains significantly lower queue backlogs than other methods, indicating better stability and lower queueing delays. This validates the theoretical queue stability guarantees provided in “Theoretical analysis”.
We also analyzed the impact of the Lyapunov parameter V on the system performance. Figure 15 shows the tradeoff between throughput and queue backlog as V varies. As V increases, the throughput improves but at the cost of larger queue backlogs. This demonstrates the [O(1/V), O(V)] utility-backlog tradeoff predicted by the theoretical analysis in Sect. 5.
To evaluate the quantum speedup, we compared the training time of our quantum-enhanced algorithm against a classical implementation. Figure 16 shows the training time as a function of the input dimension. As predicted by Theorem 3, the quantum-enhanced training achieves a quadratic speedup, with the gap widening for larger input dimensions.
Finally, we analyzed the QML framework’s adaptation to dynamic network conditions. Figure 17 shows how the offloading policy evolves over time as channel conditions change. The QML approach quickly adapts its policy to maintain high performance, outperforming static optimization approaches.
Discussion
The simulations’ outcome proves the performance improvements of the proposed QML framework for computation offloading in MEC networks. Several key insights can be drawn:
-
1.
Superior Performance: This case shows that the QML approach achieves a higher throughput, lower energy consumption, and lower queue variance than a baseline method. This is a significant accomplishment in integrating how the use of quantum neural networks along with Lyapunov optimization can lead to the development of adaptive offloading policies.
-
2.
Stability Guarantees: QML’s ability to simulate low and stable queue backlogs supports the theoretical stability promises offered by the Lyapunov optimization framework. This keeps the system stable regardless of ever-changing network conditions.
-
3.
Quantum Advantage: The training algorithm with quantum elements is significantly faster and more efficient than the existing classical approaches, particularly with high-dimensional inputs. It allows for easier work within new network conditions.
-
4.
Scalability: QML’s relative improvement increases with the network’s size, suggesting that it scales well with practical MEC scenarios.
-
5.
Adaptivity: The performance analysis has shown that QML is more efficient in time-varying conditions and channel traffic loads than static optimization.
-
6.
Tradeoff Control: The absolute value of the Lyapunov parameter V indicates the level of the tradeoff between the throughput performance achieved and the amount of queue backlogs needed to achieve it.
These results show the prospect of utilizing quantum machine learning to enhance resource management in next-generation MEC networks fundamentally. Here, quantum computing is leveraged to solve the high-dimensional optimization inherent to the MEC system and to learn flexible control strategies in light of network scenarios.
Quantum circuit implementation details
Quantum circuit architecture for QNN layers
The quantum neural network implementation consists of specialized quantum circuits designed for efficient computation offloading optimization. The architecture includes:
Quantum convolutional layer design.
-
Each quantum convolutional filter uses 4 qubits arranged in a linear topology.
-
Single-qubit rotation gates (RX, RY, RZ) are applied to each qubit for feature extraction.
-
Two-qubit controlled-Z (CZ) gates create entanglement between adjacent qubits.
-
Quantum pooling is implemented via Pauli-Z measurements averaged over qubit groups.
Gate-level implementation.
function quantumConvLayer(inputState) % Initialize parameters numQubits = 4; % Number of qubits per filter theta1 = rand(numQubits, 1); % RX rotation angles theta2 = rand(numQubits, 1); % RY rotation angles theta3 = rand(numQubits, 1); % RZ rotation angles.

end.
% Helper functions function state = applyRXGate(state, qubit, theta) % Apply RX rotation gate rotationMatrix = [cos(theta/2), -1isin(theta/2); -1isin(theta/2), cos(theta/2)]; state = applyGate(state, rotationMatrix, qubit); end.
function state = applyRYGate(state, qubit, theta) % Apply RY rotation gate rotationMatrix = [cos(theta/2), -sin(theta/2); sin(theta/2), cos(theta/2)]; state = applyGate(state, rotationMatrix, qubit); end.
function state = applyRZGate(state, qubit, theta) % Apply RZ rotation gate rotationMatrix = [exp(-1itheta/2), 0; 0, exp(1itheta/2)]; state = applyGate(state, rotationMatrix, qubit); end.
function state = applyCZGate(state, qubit1, qubit2) % Apply controlled-Z gate between two qubits czMatrix = [1 0 0 0; 0 1 0 0; 0 0 1 0; 0 0 0–1]; state = applyTwoQubitGate(state, czMatrix, qubit1, qubit2); end.
function expectation = measurePauliZ(state, qubits) % Measure Pauli-Z expectation value pauliZ = [1 0; 0–1]; expectation = computeExpectationValue(state, pauliZ, qubits); end.
function state = applyGate(state, gate, qubit) % Apply single-qubit gate dimension = length(state); identity = eye(2); fullGate = 1;

end.
Circuit optimization and error mitigation
Circuit depth optimization.
-
Parallel Execution of independent single-qubit gates.
-
Optimized CZ gate scheduling to minimize circuit depth.
-
Efficient quantum pooling implementation using ancilla qubits.
Error mitigation strategies.
-
Zero-noise extrapolation for systematic error reduction.
-
Measurement error mitigation using calibration matrices.
-
Dynamical decoupling sequences between gates.
-
Post-selection based on quantum error detection.
Quantum state preparation
The input features (queue backlogs, channel states) are encoded into quantum states using:
Table 7 compares the computational complexity of quantum and classical algorithms for solving the mobile edge computing (MEC) offloading optimization problem. As discussed in “Quantum speedup”, the quantum approach achieves significant speedups, particularly for large-scale networks. The table quantifies these advantages, showing exponential speedup for policy optimization and quadratic speedup for resource allocation. This aligns with the theoretical analysis of quantum speedup presented in Theorem 3, demonstrating the potential of quantum computing to reduce computation time for complex MEC network optimization tasks dramatically.
Measurement and classical post-processing
Measurement protocol.
-
Expectation values of Pauli observables measured in a computational basis.
-
Multiple measurement shots for statistical averaging.
-
Adaptive measurement schemes based on confidence thresholds.
Classical integration.
-
Measurement outcomes processed by classical neural network layers.
-
Efficient classical-quantum interface design.
-
Real-time feedback for dynamic circuit adjustment.
This detailed quantum circuit implementation enables the QML framework to improve performance while maintaining practical feasibility on near-term quantum devices.
Comparative performance analysis
Benchmarking against classical algorithms
Our quantum machine learning framework demonstrates significant performance improvements compared to classical approaches across multiple metrics:
Throughput performance.
-
QML achieves 30% higher throughput compared to classical Lyapunov optimization.
-
Maintains consistent performance advantage across different network loads.
-
Shows better adaptation to dynamic channel conditions.
Computational efficiency.
Table 8 summarizes the key performance improvements achieved by the proposed Quantum Machine Learning (QML) framework compared to baseline offloading methods. The table highlights significant gains in throughput, power efficiency, queue stability, and computational complexity.
As discussed in “Performance evaluation”, the QML approach achieves up to a 30% increase in average system throughput and over 20% reduction in mobile device power consumption compared to the best-performing baseline methods. Furthermore, the QML framework maintains lower average queue backlogs, indicating improved stability and reduced delays. The table also emphasizes the computational advantages of the quantum-enhanced algorithm, which provides exponential speedup for policy optimization and quadratic speedup for resource allocation compared to classical approaches. These quantum speedups enable the QML framework to efficiently handle large-scale MEC networks with high-dimensional state and action spaces.
Classical-quantum hybrid simulations
We developed a classical-quantum hybrid simulator to evaluate the performance of our proposed QML approach for computation offloading in MEC networks further. This simulator combines classical neural networks with simulated quantum circuits to comprehensively analyze our QML framework compared to purely classical baselines.
The hybrid simulator was implemented using a combination of TensorFlow for classical neural network operations and Qiskit for quantum circuit simulations. Qiskit’s quantum circuit simulator simulated the quantum layers of our QML model, while TensorFlow’s neural network modules implemented the classical layers.
Our hybrid simulator allowed us to investigate the performance of the QML approach under various network conditions and scales that would be challenging to implement on current quantum hardware. We simulated MEC networks with up to 50 mobile devices and 10 edge servers, far exceeding the capabilities of current quantum processors.
The simulation results demonstrated significant advantages of the QML approach over classical methods:
Throughput improvement
The QML framework achieved an average throughput improvement of 28.5% compared to the best classical baseline (deep reinforcement learning).
Energy Efficiency
Our approach reduced overall energy consumption by 22.3% while maintaining higher throughput, showcasing its ability to balance performance and energy efficiency.
Scalability
The QML model showed superior scalability, maintaining its performance advantage as the network size increased from 10 to 50 devices, while classical methods experienced degradation.
Adaptability
The QML approach demonstrated faster adaptation to dynamic network conditions, converging to near-optimal policies 35% faster than classical reinforcement learning methods.
Computational efficiency
Despite the overhead of quantum circuit simulation, the hybrid approach showed a 15% reduction in overall training time compared to purely classical deep learning models, indicating potential for significant speedup on actual quantum hardware.
These results highlight the potential of quantum-enhanced machine learning for optimizing complex networking problems. The hybrid simulator provided valuable insights into the scalability and performance characteristics of our QML framework, which would be challenging to obtain through other means, given the current state of quantum hardware.
Figure 18 shows the convergence behavior of a hybrid neural network with quantum-inspired layers over 3100 iterations. The upper graph displays the Mean Squared Error (MSE) starting at approximately 3.5 and rapidly decreasing to stabilize around 1.5 after 500 iterations. The lower graph shows the corresponding loss metrics, with training (blue) and validation (orange) curves demonstrating stable convergence. The training was completed on November 4, 2024, using a single GPU with a learning rate of 0.001 and achieved maximum epochs completion.
Scalability analysis
Network size scaling.
-
Performance advantage increases with network size.
-
Maintains stable queue backlogs even with a larger number of devices.
-
Quantum speedup becomes more pronounced for larger networks.
Resource Requirements.
Table 9 illustrates the effect of varying the Lyapunov control parameter V on key system performance metrics, as explored in Fig. 15. It quantifies the [O(1/V), O(V)] utility-backlog tradeoff predicted by the theoretical analysis in Sect. 5.2. We observe improved throughput and reduced power consumption as V increases but at the cost of increased queue backlogs. This table provides concrete data supporting the theoretical guarantees of the QML framework and demonstrates its ability to balance multiple competing objectives in MEC networks.
Tradeoff Analysis:
Max Throughput: 2.3201 at V = 2.7826.
Min Queue Length: 2671.1896 at V = 1.
Min Power: 0.035783 at V = 10,000.
Figure 19 illustrates how the Lyapunov parameter V affects three key system metrics in the MEC network. The top subplot shows average throughput decreasing from 2 to near 0 as V increases from 10⁰ to 10⁴. The middle subplot demonstrates the average queue length rising from approximately 2700 to 5000 units with increasing V. The bottom subplot displays average power consumption declining from around 3 to 0 units as V increases. These relationships align with the theoretical [O(1/V), O(V)] utility-backlog tradeoff, where larger V values reduce power consumption and throughput while increasing queue backlogs.
Performance under dynamic conditions
The QML framework shows superior adaptation to:
-
Channel state variations.
-
Task arrival rate fluctuations.
-
Network topology changes.
-
Resource availability variations.
The graph shown in Fig. 20 demonstrates system performance under dynamic conditions across 1000 time slots, comparing multiple execution approaches. The QML Framework (blue line) consistently outperforms the Classical Algorithm (red line) and Local Execution (green line), achieving an average performance improvement of 39.6%. The yellow dotted line represents the Task Arrival Rate, while the purple dashed line shows Channel Quality fluctuations. The QML Framework adapts effectively to varying channel conditions and task arrivals, maintaining higher throughput levels especially during peaks around slots 400–600 and 700–800, where it shows significant performance advantages over traditional approaches.
Energy efficiency analysis
Power Consumption Comparison.
-
20% reduction in overall power consumption.
-
Better energy-throughput tradeoff.
-
More efficient resource utilization.
-
Lower computational overhead.
The graph shown in Fig. 21 compares energy efficiency across different offloading methods for varying numbers of mobile devices (5 to 25). The Proposed QML method demonstrates superior performance, reaching peak efficiency of 1.29 bits/Joule with 10 devices. Local Execution shows consistent but lower efficiency around 0.6–0.7 bits/Joule. Random and Greedy Offloading methods perform similarly, with efficiencies between 0.8 and 0.95 bits/Joule. Classical DL and Lyapunov Opt methods generally show the lowest efficiency, particularly at higher device counts. The key findings indicate that the Proposed QML achieves the highest energy efficiency, with performance advantages becoming more pronounced as the number of devices increases.
These comprehensive benchmarking results demonstrate the significant advantages of our quantum-enhanced approach over classical methods, particularly in terms of throughput, scalability, and energy efficiency.
Figure 22 compares the performance scaling of the proposed quantum machine learning algorithm against classical algorithms for the MEC offloading optimization problem as the network size increases. The quantum algorithm demonstrates a significant advantage, maintaining a nearly constant runtime complexity even for large network sizes with hundreds of mobile devices and edge servers. In contrast, the runtime of classical algorithms grows exponentially with the network size, quickly becoming intractable for practical MEC deployments. This superior scaling behaviour of the quantum algorithm can be attributed to its ability to efficiently explore the exponentially large optimization space by leveraging quantum superposition and parallelism. By encoding the system state and offloading decisions into quantum states, the quantum algorithm can evaluate many possible solutions simultaneously, avoiding the need for exhaustive search. Moreover, the quantum-enhanced training procedure, which utilizes quantum matrix multiplication and quantum gradient descent, provides a quadratic speedup over classical training methods. These quantum speedups compound to enable the optimization of large-scale MEC networks that are beyond the reach of classical approaches. The results highlight the immense potential of quantum machine learning to tackle the challenges of future MEC systems and support the stringent requirements of emerging applications.
Table 10 covers the nomenclature key parameters, variables, and abbreviations used throughout the figures and analysis of the quantum-enabled mobile edge computing system performance study.
Conclusion
This paper introduced a new quantum machine learning model for stabilizing computation offloading in mobile edge computing networks using Lyapunov optimization theory. The proposed innovation uses hybrid quantum-classical neural networks with Lyapunov drift analysis to learn adaptive offloading policies that satisfy both network performance and queue stability. We provided rigorous theoretical analysis to prove the stability and near-optimal performance of the proposed framework. Extensive simulations demonstrated that our QML approach significantly outperforms conventional offloading schemes, improving throughput by up to 30% while reducing energy consumption by over 20%. The quantum-enhanced training algorithm also achieves substantial speedups over classical implementations.
The proposed framework opens up exciting new possibilities for leveraging quantum computing to enhance next-generation wireless networks and services. Future work could extend the approach to other network control problems, incorporating more advanced quantum machine learning models, and experimental implementation on near-term quantum devices. With the development of quantum hardware, we believe that algorithms enhanced by quantum will be used to improve the networking algorithms used in highly complicated systems. This work is a step in the right direction towards achieving the vision of quantum technologies for the next generation of mobile edge computing and beyond. Extending the quantum machine learning approach to other network control problems, developing more advanced quantum machine learning models, and implementing the framework on near-term quantum devices to validate the theoretical performance advantages in real-world scenario study.
Data availability
The datasets used and/or analyzed during the current study are available from the corresponding author upon reasonable request.
References
Mao, Y., You, C., Zhang, J., Huang, K. & Letaief, K. B. A Survey on Mobile Edge Computing: the communication perspective. IEEE Commun. Surv. Tutorials. 19 (4), 2322–2358 (2017).
Neely, M. Stochastic Network Optimization with Application to Communication and Queueing Systems. Synth. Lect. Commun. Netw. 3(1), 1–211 (2010).
Biamonte, J. et al. Quantum Mach. Learn. Nat., 549, 7671, 195–202, (2017).
Arute, F. et al. Quantum Supremacy Using a Programmable Superconducting Processor. Nature 574(7779), 505–510 (2019).
Mao, Y., Zhang, J. & Letaief, K. B. Dynamic computation offloading for Mobile-Edge Computing with Energy Harvesting devices. IEEE J. Sel. Areas Commun. 34 (12), 3590–3605 (2016).
Schuld, M., Sinayskiy, I. & Petruccione, F. An Introduction to Quantum Machine Learning. Contemp. Phys. 56(2), 172–185 (2015).
Havlíček, V. et al. Supervised Learning with Quantum-Enhanced Feature Spaces. Nature 567(7747), 209–212 (2019).
Beer, K. et al. Training Deep Quantum Neural Networks. Nat. Commun. 11(1), 1–6 (2020).
Neely, M. Energy-Aware Wireless Scheduling with Near optimal backlog and Convergence Time Tradeoffs. IEEE/ACM Trans. Network. 24(4), 2223–2236 (2016).
Aaronson, S. Read the Fine Print, Nature Physics, vol. 11, no. 4, pp. 291–293, (2015).
Cong, I., Choi, S. & Lukin, M. D. Quantum convolutional neural networks. Nat. Phys. 15, 1273–1278. https://doi.org/10.1038/s41567-019-0648-8 (2019).
Bravo, R. A., Najafi, K., Gao, X. & Yelin, S. F. Quantum Reservoir Computing Using Arrays of Rydberg Atoms. PRX Quantum. 3 (3). https://doi.org/10.1103/prxquantum.3.030325 (2022).
Liu, Z., Shen, P., Li, W., Duan, L. & Deng, D. Quantum capsule networks. Quantum Sci. Technol. 8 (1), 015016. https://doi.org/10.1088/2058-9565/aca55d (2022).
Lu, Z., Shen, P. & Deng, D. Markovian Quantum Neuroevolution for Machine Learning. Phys. Rev. Appl. 16 (4). https://doi.org/10.1103/physrevapplied.16.044039 (2021).
Li, W., Lu, S. & Deng, D. L. Quantum federated learning through blind quantum computing. Sci. China Phys. Mech. Astron. 64, 100312. https://doi.org/10.1007/s11433-021-1753-3 (2021).
Shen, P., Jiang, W., Li, W., Lu, Z. & Deng, D. Adversarial learning in quantum artificial intelligence. Acta Phys. Sinica. 70 (14), 140302. https://doi.org/10.7498/aps.70.20210789 (2021).
Yu, L. et al. Expressibility-induced concentration of Quantum neural tangent kernels. Rep. Prog. Phys. https://doi.org/10.1088/1361-6633/ad82cf (2024).
Zhang, C. et al. Quantum continual learning on a programmable superconducting processor. arXiv (Cornell University). (2024). https://doi.org/10.48550/arxiv.2409.09729
Chen, M., Liu, A., Xiong, N. N., Song, H. & Leung, V. C. M. SGPL: An Intelligent Game-based Secure Collaborative Communication Scheme for Metaverse over 5G and beyond networks. IEEE J. Sel. Areas Commun. 42 (3), 767–782. https://doi.org/10.1109/jsac.2023.3345403 (2023).
Zhou, X., Yang, J., Li, Y., Li, S. & Su, Z. Deep reinforcement learning-based resource scheduling for energy optimization and load balancing in SDN-driven edge computing. Comput. Commun. 226–227. https://doi.org/10.1016/j.comcom.2024.107925 (2024).
Hou, H., Jawaddi, S. N. A. & Ismail, A. Energy efficient task scheduling based on deep reinforcement learning in cloud environment: a specialized review. Future Generation Comput. Syst. 151, 214–231. https://doi.org/10.1016/j.future.2023.10.002 (2023).
Mei, J., Dai, L., Tong, Z., Zhang, L. & Li, K. Lyapunov optimized energy-efficient dynamic offloading with queue length constraints. J. Syst. Architect. 143, 102979. https://doi.org/10.1016/j.sysarc.2023.102979 (2023).
Huang, P., Wang, Y., Wang, K. & Zhang, Q. Combining Lyapunov optimization with evolutionary transfer optimization for Long-Term Energy Minimization in IRS-Aided communications. IEEE Trans. Cybernetics. 53 (4), 2647–2657. https://doi.org/10.1109/tcyb.2022.3168839 (2022).
Sadilek, A. et al. Privacy-first health research with federated learning. Npj Digit. Med. 4, 132. https://doi.org/10.1038/s41746-021-00489-2 (2021).
Zhang, X., Kang, Y., Chen, K., Fan, L. & Yang, Q. Trading off privacy, Utility, and efficiency in Federated Learning. ACM Trans. Intell. Syst. Technol. 14 (6), 1–32. https://doi.org/10.1145/3595185 (2023).
Huang, X., Leng, S., Maharjan, S. & Zhang, Y. Multi-agent deep reinforcement learning for Computation Offloading and Interference Coordination in Small Cell Networks. IEEE Trans. Veh. Technol. 70 (9), 9282–9293. https://doi.org/10.1109/tvt.2021.3096928 (2021).
Arowoiya, V. A., Moehler, R. C. & Fang, Y. Digital twin technology for thermal comfort and energy efficiency in buildings: a state-of-the-art and future directions. Energy Built Environ. https://doi.org/10.1016/j.enbenv.2023.05.004 (2023).
Bi, S., Huang, L., Wang, H. & Zhang, Y. A. Lyapunov-guided deep reinforcement learning for stable online computation offloading in Mobile-Edge Computing Networks. IEEE Trans. Wireless Commun. 20 (11), 7519–7537. https://doi.org/10.1109/twc.2021.3085319 (2021).
Lal, R. et al. AI-based optimization of EM radiation estimates from GSM base stations using traffic data. Discov Appl. Sci. 6, 655. https://doi.org/10.1007/s42452-024-06395-y (2024).
Belgacem, A. et al. Efficient dynamic resource allocation method for cloud computing environment. Cluster Comput. 23, 2871–2889. https://doi.org/10.1007/s10586-020-03053-x (2020).
Zhou, Z., Ye, F., Li, Y., Geng, X. & Zhang, S. Blockchain-assisted computation offloading collaboration: a hierarchical game to fortify IoT security and resilience. Internet Things. 23, 100880. https://doi.org/10.1016/j.iot.2023.100880 (2023).
Song, J., Song, Q., Wang, Y. & Lin, P. Energy–Delay Tradeoff in Adaptive Cooperative Caching for Energy-Harvesting Ultradense Networks. IEEE Trans. Comput. Social Syst. 9 (1), 218–229. https://doi.org/10.1109/tcss.2021.3097335 (2021).
Chen, Y., Li, Z., Yang, B., Nai, K. & Li, K. A Stackelberg game approach to multiple resources allocation and pricing in mobile edge computing. Future Generation Comput. Syst. 108, 273–287. https://doi.org/10.1016/j.future.2020.02.045 (2020).
Cerezo, M. et al. Challenges and opportunities in quantum machine learning. Nat. Comput. Sci. 2, 567–576. https://doi.org/10.1038/s43588-022-00311-3 (2022).
Peral-García, D., Cruz-Benito, J. & García-Peñalvo, F. J. Systematic literature review: Quantum machine learning and its applications. Comput. Sci. Rev. 51, 100619. https://doi.org/10.1016/j.cosrev.2024.100619 (2024).
Zeguendry, A., Jarir, Z. & Quafafou, M. Quantum Machine Learning: a review and Case studies. Entropy 25, 287. https://doi.org/10.3390/e25020287 (2023).
M, S. Machine learning and quantum computing for 5G/6G communication networks - a survey. Int. J. Intell. Networks. 3, 197–203. https://doi.org/10.1016/j.ijin.2022.11.004 (2022).
Wang, M., Zhang, Y., He, X. & Yu, S. Joint scheduling and offloading of computational tasks with time dependency under edge computing networks. Simul. Model. Pract. Theory. 129, 102824. https://doi.org/10.1016/j.simpat.2023.102824 (2023).
Alfakih, T., Hassan, M. M., Gumaei, A., Savaglio, C. & Fortino, G. Task Offloading and Resource Allocation for Mobile Edge Computing by deep reinforcement learning based on SARSA. IEEE Access. 8, 54074–54084. https://doi.org/10.1109/access.2020.2981434 (2020).
Sadatdiynov, K. et al. A review of optimization methods for computation offloading in edge computing networks. Digit. Commun. Networks. 9 (2), 450–461. https://doi.org/10.1016/j.dcan.2022.03.003 (2022).
Bhat, R. V., Vaze, R. & Motani, M. Throughput maximization with an average age of Information Constraint in Fading channels. IEEE Trans. Wireless Commun. 20 (1), 481–494. https://doi.org/10.1109/twc.2020.3025630 (2020).
Author information
Authors and Affiliations
Contributions
V.V. and D.K.N. conceived the study, designed the quantum machine learning framework, and wrote the manuscript. V.S. and V.K.S. developed the theoretical analysis and proofs, while A.V. conducted simulations and prepared figures. D.R.S. provided guidance on quantum circuit implementations. Specifically, V.V. and D.K.N. made substantial contributions to the conception, design, and drafting; V.S. and V.K.S. contributed to theory and proofs; A.V. acquired and analyzed data; and D.R.S. provided quantum circuit expertise. All authors discussed results, reviewed the manuscript, approved the submitted version, and agreed to be accountable for their own contributions and the work’s integrity.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Verma, V.R., Nishad, D.K., Sharma, V. et al. Quantum machine learning for Lyapunov-stabilized computation offloading in next-generation MEC networks. Sci Rep 15, 405 (2025). https://doi.org/10.1038/s41598-024-84441-w
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-024-84441-w