1 |
SCHWARTING W, ALONSO-MORA J, RUS D. Planning and decision-making for autonomous vehicles [J]. Annual Review of Control, Robotics, and Autonomous Systems, 2018, 1(1): 187-210.
|
2 |
LIKMETA A, METELLI A M, TIRINZONI A, et al. Combining reinforcement learning with rule-based controllers for transparent and general decision-making in autonomous driving [J]. Robotics and Autonomous Systems, 2020, 131: 103568.
|
3 |
ZHAO L, ICHISE R, SASAKI Y, et al. Fast decision making using ontology-based knowledge base[C]. Proceedings of the 2016 IEEE Intelligent Vehicles Symposium (IV), F 19-22 June 2016.
|
4 |
VACEK S, GINDELE T, ZOLLNER J M, et al. Using case-based reasoning for autonomous vehicle guidance[C]. Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, F 29 Oct.-2 Nov. 2007.
|
5 |
CHEN C, SEFF A, KORNHAUSER A, et al. DeepDriving: learning affordance for direct perception in autonomous driving[C]. Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), F 7-13 Dec. 2015.
|
6 |
DESJARDINS C, CHAIB-DRAA B. Cooperative adaptive cruise control: a reinforcement learning approach [J]. IEEE Transactions on Intelligent Transportation Systems, 2011, 12(4): 1248-1260.
|
7 |
SHU H, LIU T, MU X, et al. Driving tasks transfer using deep reinforcement learning for decision-making of autonomous vehicles in unsignalized intersection [J]. IEEE Transactions on Vehicular Technology, 2022, 71(1): 41-52.
|
8 |
MNIH V, KAVUKCUOGLU K, SILVER D, et al. Human-level control through deep reinforcement learning [J]. Nature, 2015, 518(7540): 529-533.
|
9 |
VITELLI M, NAYEBI A. Carma: a deep reinforcement learning approach to autonomous driving [D]. Tech Rep Stanford University, Tech Rep, 2016.
|
10 |
FU Y, LI C, YU F R, et al. Hybrid autonomous driving guidance strategy combining deep reinforcement learning and expert system [J]. IEEE Transactions on Intelligent Transportation Systems, 2022, 23(8): 11273-11286.
|
11 |
KAELBLING L P, LITTMAN M L, CASSANDRA A R. Planning and acting in partially observable stochastic domains [J]. Artificial Intelligence, 1998, 101(1): 99-134.
|
12 |
CAI P, LUO Y, HSU D, et al. HyP-DESPOT: a hybrid parallel algorithm for online planning under uncertainty [J]. The International Journal of Robotics Research, 2021, 40(2-3): 558-573.
|
13 |
SILVER D, VENESS J. Monte-Carlo planning in large POMDPs [J]. Advances in Neural Information Processing Systems, 2010, 23.
|
14 |
YE N, SOMANI A, HSU D, et al. DESPOT: online POMDP planning with regularization [J]. Journal of Artificial Intelligence Research, 2017, 58: 231-266.
|
15 |
KURNIAWATI H, YADAV V. An online POMDP solver for uncertainty planning in dynamic environment [M]. Robotics Research. Springer, 2016: 611-629.
|
16 |
HUBMANN C, SCHULZ J, XU G, et al. A belief state planner for interactive merge maneuvers in congested traffic[C]. Proceedings of the 2018 21st International Conference on Intelligent Transportation Systems (ITSC), F 4-7 Nov. 2018.
|
17 |
BAI H, CAI S, YE N, et al. Intention-aware online POMDP planning for autonomous driving in a crowd[C]. Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), F 26-30 May 2015.
|
18 |
HUBMANN C, BECKER M, ALTHOFF D, et al. Decision making for autonomous driving considering interaction and uncertain prediction of surrounding vehicles[C]. Proceedings of the 2017 IEEE Intelligent Vehicles Symposium (IV), F 11-14 June 2017.
|
19 |
HUBMANN C, SCHULZ J, BECKER M, et al. Automated driving in uncertain environments: planning with interaction and uncertain maneuver prediction [J]. IEEE Transactions on Intelligent Vehicles, 2018, 3(1): 5-17.
|
20 |
GALCERAN E, CUNNINGHAM A G, EUSTICE R M, et al. Multipolicy decision-making for autonomous driving via changepoint-based behavior prediction: theory and experiment [J]. Autonomous Robots, 2017, 41(6): 1367-1382.
|
21 |
CUNNINGHAM A G, GALCERAN E, EUSTICE R M, et al. MPDM: multipolicy decision-making in dynamic, uncertain environments for autonomous driving[C]. Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), F 26-30 May 2015.
|
22 |
NISHI T, DOSHI P, PROKHOROV D. Merging in congested freeway traffic using multipolicy decision making and passive actor-critic learning [J]. IEEE Transactions on Intelligent Vehicles, 2019, 4(2): 287-297.
|
23 |
DING W, ZHANG L, CHEN J, et al. EPSILON: an efficient planning system for automated vehicles in highly interactive environments [J]. IEEE Transactions on Robotics, 2022, 38(2): 1118-1138.
|
24 |
ZHANG L, DING W, CHEN J, et al. Efficient uncertainty-aware decision-making for automated driving using guided branching[C]. Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), F 31 May-31 Aug. 2020.
|
25 |
WATTS D J, STROGATZ S H. Collective dynamics of ‘small-world’networks [J]. Nature, 1998, 393(6684): 440-442.
|
26 |
BOCCALETTI S, LATORA V, MORENO Y, et al. Complex networks: structure and dynamics [J]. Physics Reports, 2006, 424(4-5): 175-308.
|
27 |
LIU Y Y, SLOTINE J J, BARABáSI A L. Controllability of complex networks [J]. Nature, 2011, 473(7346): 167-173.
|
28 |
YUAN Z, ZHAO C, DI Z, et al. Exact controllability of complex networks [J]. Nature Communications, 2013, 4(1): 2447.
|
29 |
BAGGIO G, BASSETT D S, PASQUALETTI F. Data-driven control of complex networks [J]. Nature Communications, 2021, 12(1): 1429.
|
30 |
LI A, CORNELIUS S P, LIU Y Y, et al. The fundamental advantages of temporal networks [J]. Science, 2017, 358(6366): 1042-1046.
|
31 |
UNICOMB S, IñIGUEZ G, GLEESON J P, et al. Dynamics of cascades on burstiness-controlled temporal networks [J]. Nature Communications, 2021, 12(1): 133.
|
32 |
STROGATZ S H. Exploring complex networks [J]. Nature, 2001, 410(6825): 268-276.
|
33 |
ZOU Y, DONNER R V, MARWAN N, et al. Complex network approaches to nonlinear time series analysis [J]. Physics Reports, 2019, 787:1-97.
|
34 |
TREIBER M, KESTING A. Traffic flow dynamics [J]. Traffic Flow Dynamics: Data, Models Simulation, Springer-Verlag Berlin Heidelberg, 2013, 983-1000.
|
35 |
COULTER R C. Implementation of the pure pursuit path tracking algorithm[M]. Carnegie Mellon University, The Robotics Institute, 1992.
|