TY - GEN
T1 - Towards Task Number Adaptive Offloading in MEC Systems
T2 - 101st IEEE Vehicular Technology Conference, VTC 2025-Spring 2025
AU - Xie, Yiping
AU - Zhang, Fan
AU - Fu, Yaru
AU - Xu, Chao
AU - Quek, Tony Q.S.
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - In a practical Mobile Edge Computing (MEC) system, the stochastic arrival and departure of heterogeneous Mobile Devices (MDs) can cause fluctuations in the number of generated tasks to be scheduled over time, with the generated tasks differing in data size, complexity, and delay constraint. In this work, we consider a dynamic Task Offloading (TO) problem aiming at maximizing the long-term average system utility jointly defined by task completion and energy consumption in the MEC system, where the number of MDs to be served varies over time, and the processing of generated tasks may span multiple time slots. Particularly, we first transform the TO problem into a Markov Decision Process (MDP) with time-varying state and action spaces, which cannot be effectively solved by conventional Deep Reinforcement Learning (DRL) algorithms. Then, we propose a state and action spaces adaptive DRL algorithm to efficiently solve the formulated MDP by leveraging the Transformer model. Finally, simulation results demonstrate the superiority of our proposed algorithm over baseline algorithms and emphasize the limitation of the conventional DRL algorithm in handling time-varying state and action spaces.
AB - In a practical Mobile Edge Computing (MEC) system, the stochastic arrival and departure of heterogeneous Mobile Devices (MDs) can cause fluctuations in the number of generated tasks to be scheduled over time, with the generated tasks differing in data size, complexity, and delay constraint. In this work, we consider a dynamic Task Offloading (TO) problem aiming at maximizing the long-term average system utility jointly defined by task completion and energy consumption in the MEC system, where the number of MDs to be served varies over time, and the processing of generated tasks may span multiple time slots. Particularly, we first transform the TO problem into a Markov Decision Process (MDP) with time-varying state and action spaces, which cannot be effectively solved by conventional Deep Reinforcement Learning (DRL) algorithms. Then, we propose a state and action spaces adaptive DRL algorithm to efficiently solve the formulated MDP by leveraging the Transformer model. Finally, simulation results demonstrate the superiority of our proposed algorithm over baseline algorithms and emphasize the limitation of the conventional DRL algorithm in handling time-varying state and action spaces.
KW - deep reinforcement learning
KW - edge computing
KW - energy-efficient offloading
KW - time-varying state and action spaces
UR - https://www.scopus.com/pages/publications/105019038334
U2 - 10.1109/VTC2025-Spring65109.2025.11174688
DO - 10.1109/VTC2025-Spring65109.2025.11174688
M3 - Conference contribution
AN - SCOPUS:105019038334
T3 - IEEE Vehicular Technology Conference
BT - 2025 IEEE 101st Vehicular Technology Conference, VTC 2025-Spring 2025 - Proceedings
Y2 - 17 June 2025 through 20 June 2025
ER -