Automotive Engineering ›› 2022, Vol. 44 ›› Issue (7): 969-975.doi: 10.19562/j.chinasae.qcgc.2022.07.003
Special Issue: 智能网联汽车技术专题-规划&控制2022年
Previous Articles Next Articles
Zhenhai Gao,Xiangtong Yan,Fei Gao()
Received:
2022-01-05
Revised:
2022-02-16
Online:
2022-07-25
Published:
2022-07-20
Contact:
Fei Gao
E-mail:gaofei123284123@jlu.edu.cn
Zhenhai Gao,Xiangtong Yan,Fei Gao. A Decision-making Method for Longitudinal Autonomous Driving Based on Inverse Reinforcement Learning[J].Automotive Engineering, 2022, 44(7): 969-975.
1 | 邱锡鹏.神经网络与深度学习[M].北京:机械工业出版社,2020: 329. |
QIU X P. Neural networks and deep learning[M]. Beijing: China Machine Press, 2020: 329. | |
2 | CHARLES D, BRAHIM C. Cooperative adaptive cruise control: a reinforcement learning approach[J].IEEE Transactions on Intelligent Transportation Systems, 2011, 12(4): 1248-1260. |
3 | 高振海,孙天骏,何 磊.汽车纵向自动驾驶的因果推理型决策[J].吉林大学学报(工学版),2019,49(5):1392-1404. |
GAO Z H, SUN T J, HE L. Causal reasoning decision⁃making for vehicle longitudinal autonomous driving[J]. Journal of Jilin University (Engineering and Technology Edition), 2019, 49(5): 1392-1404. | |
4 | YE Y J, ZHANG X H, SUN J. Automated vehicle's behavior decision making using deep reinforcement learning and high-fidelity simulation environment[J].Transportation Research: Part C, 2019, 107: 155-170. |
5 | ZHU M X, WANG Y H, PU Z Y, et al. Safe, efficient, and comfortable velocity control based on reinforcement learning for autonomous driving[J].Transportation Research Part C: Emerging Technologies, 2020, 117:102662. |
6 | 郭应时, 蒋拯民, 白 艳,等.无人驾驶汽车路径跟踪控制方法拟人程度研究[J]. 中国公路学报,2018,31(8):189-196. |
GUO Y S, JIANG Z M, BAI Y, et al. Investigation of humanoid level of path tracking methods based on autonomous vehicles[J]. China Journal of Highway and Transport, 2018, 31(8): 189-196. | |
7 | BLESA W, EBOSA J E, GRAAF B, et al. Motion sickness: only one provocative conflict[J]. Brain Research Bulletin, 1998, 47(5):481-487. |
8 | PAN F, BAO H.Reinforcement learning model with a reward function based on human driving characteristics[C]. Proceedings of the 15th International Conference on Computational Intelligence and Security, IEEE, 2019: 225-229. |
9 | GU T, DOLAN J M. Toward human-like motion planning in urban environments[C]. 2014 IEEE Intelligent Vehicles Symposium Proceedings, IEEE, 2014: 350-355. |
10 | LEFÈVRE S, CARVALHO A, BORRELLI F. Autonomous car following: a learning-based approach[C]. 2015 IEEE Intelligent Vehicles Symposium (IV), IEEE, 2015: 920–926. |
11 | KUDERER M S, BURGARD W. Learning driving styles for autonomous vehicles from demonstration[C]. IEEE International Conference on Robotics and Automation (ICRA), 2015:2641–2646. |
12 | ZOU Q J, XIONG K, HOU Y L. An end-to-end learning of driving strategies based on DDPG and imitation learning[C]. 2020 Chinese Control and Decision Conference (CCDC), IEEE, 2020: 3190-3195. |
13 | ZHU M X, WANG X S, WANG Y H, et al. Human-like autonomous car-following model with deep reinforcement learning[J]. Transportation Research: Part C,2018, 97: 348-368. |
14 | CHEN X, ZHAI Y, LU C, et al. A learning model for personalized adaptive cruise control[C]. 2017 IEEE Intelligent Vehicles Symposium (IV), IEEE, 2017: 379-384. |
15 | 刘旖菲. 逆向强化学习研究概述[J]. 电脑知识与技术, 2021,17(15): 190-191. |
LIU Y F. An overview of inverse reinforcement learning[J]. Computer Knowledge and Technology, 2021, 17(15): 190-191. | |
16 | GAO H B, SHI G Y, XIE G T, et al. Car-following method based on inverse reinforcement learning for autonomous vehicle decision-making[J]. International Journal of Advanced Robotic Systems, 2018, 15(6): 1-11. |
17 | NG A Y, RUSSELL S J. Algorithms for inverse reinforcement learning[C]. Proceedings of 17th International Conference on Machine Learning, Morgan Kaufmann, 2000: 663-670. |
18 | 冯 超. 强化学习精要:核心算法与TensorFlow实现[M]. 北京: 电子工业出版社,2018:329-371. |
FENG C. Essentials of reinforcement learning: core algorithm and Tensorflow implementation[M]. Beijing: Publishing House of Electronics Industry, 2018:329-371. | |
19 | 周志华. 机器学习[ M].北京:清华大学出版社, 2015:390-392. |
ZHOU Z H. Machine learning[M]. Beijing: Tsinghua University Publishing House, 2015:390-392. | |
20 | YOU C X, LU J B, FILEV D, et al. Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning[J]. Robotics and Autonomous Systems, 2019, 144: 1-18. |
21 | 唐明弘,高振海,宋学伟. 基于逆向强化学习的ACC拟人化决策算法研究[C].2020中国汽车工程学会年会,2020:125-131. |
TANG M H, GAO Z H, SONG X W, et al. Research on the human-like decision algorithm of ACC based on inverse reinforcement learning [C]. 2020 China Society of Automotive Engineering Congress, 2020:125-131. | |
22 | XIA C, KARMEL A E. Neural inverse reinforcement learning in autonomous navigation[J]. Robotics and Autonomous Systems, 2016, 84:1-14. |
23 | ABBEEL P, NG A Y. Apprenticeship learning via inverse reinforcement learning[C]. International Conference on Machine Learning,2004:1-8. |
24 | 高振海,闫相同,高菲. 驾驶员DDPG汽车纵向自动驾驶决策方法[J]. 汽车工程,2021,43(12): 1737-1744. |
GAO Z H, YAN X T, GAO F. A driver-like decision-making method for longitudinal autonomous driving based on DDPG [J]. Automotive Engineering, 2021,43(12): 1737-1744. |
[1] | Xiaocong Zhao,Shiyu Fang,Zirui Li,Jian Sun. Extraction and Application of Key Utility Term for Social Driving Interaction [J]. Automotive Engineering, 2024, 46(2): 230-240. |
[2] | Yanli Ma, Qin Qin, Fangqi Dong, Yining Lou. Takeover Risk Assessment Model Based on Risk Field Theory Under Different Cognitive Secondary Tasks [J]. Automotive Engineering, 2024, 46(1): 9-17. |
[3] | Weiguo Liu,Zhiyu Xiang,Weiping Liu,Daoxin Qi,Zixu Wang. Research on Vehicle Control Algorithm Based on Distributed Reinforcement Learning [J]. Automotive Engineering, 2023, 45(9): 1637-1645. |
[4] | Ming Wang,Xiaolin Tang,Kai Yang,Guofa Li,Xiaosong Hu. A Motion Planning Method for Autonomous Vehicles Considering Prediction Risk [J]. Automotive Engineering, 2023, 45(8): 1362-1372. |
[5] | Dongyu Zhao, Shuen Zhao. Autonomous Driving 3D Object Detection Based on Cascade YOLOv7 [J]. Automotive Engineering, 2023, 45(7): 1112-1122. |
[6] | Jiahao Zhao,Zhiquan Qi,Zhifeng Qi,Hao Wang,Lei He. Calculation of Heading Angle of Parallel Large Vehicle Based on Tire Feature Points [J]. Automotive Engineering, 2023, 45(6): 1031-1039. |
[7] | Lisheng Jin,Guangde Han,Xianyi Xie,Baicang Guo,Guofeng Liu,Wentao Zhu. Review of Autonomous Driving Decision-Making Research Based on Reinforcement Learning [J]. Automotive Engineering, 2023, 45(4): 527-540. |
[8] | Qingyang Huang,Xiaoping Jin,Yikang Zhang. Analysis of Drivers' Driving Posture Change Rule Under the Condition of Automatic Driving Level Improvement [J]. Automotive Engineering, 2023, 45(3): 382-392. |
[9] | Zhengfa Liu,Ya Wu,Peigen Liu,Rongqi Gu,Guang Chen. Cross-Domain Object Detection for Intelligent Driving Based on Joint Distribution Matching of Features and Labels [J]. Automotive Engineering, 2023, 45(11): 2082-2091. |
[10] | Jie Li,Xiaodong Wu,Min Xu,Yonggang Liu. Reinforcement Learning Based Multi-objective Eco-driving Strategy in Urban Scenarios [J]. Automotive Engineering, 2023, 45(10): 1791-1802. |
[11] | Chunyang Qi,Chuanxue Song,Shixin Song,Liqiang Jin,Da Wang,Feng Xiao. Research on Energy Management Strategy for Hybrid Electric Vehicles Based on Inverse Reinforcement Learning [J]. Automotive Engineering, 2023, 45(10): 1954-1964. |
[12] | Pulei Xu,Yingfeng Cai,Yubo Lian,Xiaoqiang Sun,Hai Wang,Long Chen,Yilin Zhong. AFS/DYC Coordinated Control of Intelligent Vehicles Based on Improved Hierarchical Extensibility Theory [J]. Automotive Engineering, 2023, 45(1): 20-31. |
[13] | Runhui Huang,Likun Hu,Mingfang Su,Daye Xu,Aoran Chen. Semantic Segmentation Method of LiDAR Point Cloud Based on 3D Conical Grid [J]. Automotive Engineering, 2022, 44(8): 1173-1182. |
[14] | Jiangkun Li,Weiwen Deng,Bingtao Ren,Wenqi Wang,Juan Ding. Automatic Driving Edge Scene Generation Method Based on Scene Dynamics and Reinforcement Learning [J]. Automotive Engineering, 2022, 44(7): 976-986. |
[15] | Dafang Wang,Jingdong Du,Jiang Cao,Mei Zhang,Gang Zhao. Research on Style Transfer Network for Autonomous Driving Data Generation [J]. Automotive Engineering, 2022, 44(5): 684-690. |