Administrator by China Associction for Science and Technology
Sponsored by China Society of Automotive Engineers
Published by AUTO FAN Magazine Co. Ltd.

Automotive Engineering ›› 2023, Vol. 45 ›› Issue (9): 1637-1645.doi: 10.19562/j.chinasae.qcgc.2023.09.012

Special Issue: 智能网联汽车技术专题-控制2023年

Previous Articles     Next Articles

Research on Vehicle Control Algorithm Based on Distributed Reinforcement Learning

Weiguo Liu1,2(),Zhiyu Xiang1,Weiping Liu2,Daoxin Qi2,Zixu Wang2   

  1. 1.School of Information and Electronic Engineering,Zhejiang University,Hangzhou  310058
    2.National Innovation Center of Intelligent and Connected Vehicles,Beijing  100160
  • Received:2023-04-18 Revised:2023-06-23 Online:2023-09-25 Published:2023-09-23
  • Contact: Weiguo Liu E-mail:liuweiguo@china-icv.cn

Abstract:

The development of end-to-end autonomous driving algorithms has become a hot topic in current autonomous driving technology research and development. Classic reinforcement learning algorithms leverage information such as vehicle state and environmental feedback to train the vehicle for driving, through trial-and-error learning to obtain the best strategy, so as to achieve the development of end-to-end autonomous driving algorithms. However, there is still the problem of low development efficiency. The article proposes an asynchronous distributed reinforcement learning framework to address the inefficiency and high complexity problems in training RL algorithms in virtual simulation environment, establishes intra and inter process multi-agent parallel Soft Actor-Critic (SAC) distributed training framework on the Carla simulator to accelerate online RL training. Additionally, to achieve rapid model training and deployment, the article proposes a distributed model training and deployment system architecture based on Cloud-OTA, which mainly consists of an Over-the-Air Technology (OTA) platform, a cloud-based distributed training platform, and an on-vehicle computing platform. On this basis, the paper establishes an Autoware-Carla integrated validation framework based on ROS to improve model reusability and reduce migration and deployment cost. The experimental results show that compared with various mainstream autonomous driving methods, the method proposed in this paper has a faster training speed qualitatively, which can effectively cope with dense traffic flow and improve the adaptability of end-to-end autonomous driving strategies to unknown scenes, and reduce the time and resources required for experimentation in actual environment.

Key words: reinforcement learning, distributed system, multi-agent, autonomous driving, Carla, vehicle control