Administrator by China Associction for Science and Technology
Sponsored by China Society of Automotive Engineers
Published by AUTO FAN Magazine Co. Ltd.

Automotive Engineering ›› 2025, Vol. 47 ›› Issue (1): 35-43.doi: 10.19562/j.chinasae.qcgc.2025.01.004

Previous Articles     Next Articles

ICV Task Offloading and Resource Allocation Based on Hybrid Deep Reinforcement Learning

Jiahui Liu1,2,Yuan Zou1,2(),Wei Sun1,2,Yihao Meng1,2,Xiaoran Lu1,2,Yuanyuan Li1,2   

  1. 1.School of Mechanical Engineering,Beijing Institute of Technology,Beijing 100081
    2.National Engineering Research Center for Electric Vehicles,Beijing Institute of Technology,Beijing 100081
  • Received:2024-05-20 Revised:2024-07-25 Online:2025-01-25 Published:2025-01-17
  • Contact: Yuan Zou E-mail:zouyuanbit@vip.163.com

Abstract:

With the development of Intelligent Connected Vehicle (ICV) technology, ICVs with limited computing resources face the problem of significantly increased computational demand. ICVs can offload tasks to Mobile Edge Computing (MEC) servers via Roadside Units (RSU). However, the dynamic and complex nature of vehicular networks makes task offloading and resource allocation highly challenging. In this paper, it is proposed to minimize task computing energy consumption by controlling task offloading decision, communication power, and computing resource allocation under environmental and resource constraints. To address the coexistence of discrete and continuous control variables in the problem, a Hybrid Deep Reinforcement Learning (HDRL) algorithm is designed. The algorithm employs the Double Deep Q-Network (DDQN) to generate task offloading decisions and the Deep Deterministic Policy Gradient (DDPG) to determine communication power and MEC resource allocation. Furthermore, an Improved Prioritized Experience Replay (IPER) mechanism is integrated to evaluate and select actions, outputting the optimal strategy. Simulation results show that the method achieves faster and more stable decision convergence than comparative algorithms, minimizes the energy consumption for task computation offloading, and effectively adapts to changes in the number of ICVs and task sizes, demonstrating high real-time performance and excellent environmental adaptability.

Key words: mobile edge computing, deep reinforcement learning, task offloading, resource allocation, priority experience replay