TY - GEN
T1 - Combining Reinforcement Learning with Model Predictive Control for On-Ramp Merging
AU - Lubars, Joseph
AU - Gupta, Harsh
AU - Chinchali, Sandeep
AU - Li, Liyun
AU - Raja, Adnan
AU - Srikant, R.
AU - Wu, Xinzhou
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021/9/19
Y1 - 2021/9/19
N2 - We consider the problem of designing an algorithm to allow a car to autonomously merge on to a highway from an on-ramp. Two broad classes of techniques have been proposed to solve motion planning problems in autonomous driving: Model Predictive Control (MPC) and Reinforcement Learning (RL). In this paper, we first establish the strengths and weaknesses of state-of-the-art MPC and RL-based techniques through simulations. We show that the performance of the RL agent is worse than that of the MPC solution from the perspective of safety and robustness to out-of-distribution traffic patterns, i.e., traffic patterns which were not seen by the RL agent during training. On the other hand, the performance of the RL agent is better than that of the MPC solution when it comes to efficiency and passenger comfort. We subsequently present an algorithm which blends the model-free RL agent with the MPC solution and show that it provides better tradeoffs between all metrics - passenger comfort, efficiency, crash rate and robustness.
AB - We consider the problem of designing an algorithm to allow a car to autonomously merge on to a highway from an on-ramp. Two broad classes of techniques have been proposed to solve motion planning problems in autonomous driving: Model Predictive Control (MPC) and Reinforcement Learning (RL). In this paper, we first establish the strengths and weaknesses of state-of-the-art MPC and RL-based techniques through simulations. We show that the performance of the RL agent is worse than that of the MPC solution from the perspective of safety and robustness to out-of-distribution traffic patterns, i.e., traffic patterns which were not seen by the RL agent during training. On the other hand, the performance of the RL agent is better than that of the MPC solution when it comes to efficiency and passenger comfort. We subsequently present an algorithm which blends the model-free RL agent with the MPC solution and show that it provides better tradeoffs between all metrics - passenger comfort, efficiency, crash rate and robustness.
UR - http://www.scopus.com/inward/record.url?scp=85118438345&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85118438345&partnerID=8YFLogxK
U2 - 10.1109/ITSC48978.2021.9564954
DO - 10.1109/ITSC48978.2021.9564954
M3 - Conference contribution
AN - SCOPUS:85118438345
T3 - IEEE Conference on Intelligent Transportation Systems, Proceedings, ITSC
SP - 942
EP - 947
BT - 2021 IEEE International Intelligent Transportation Systems Conference, ITSC 2021
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2021 IEEE International Intelligent Transportation Systems Conference, ITSC 2021
Y2 - 19 September 2021 through 22 September 2021
ER -