期刊文献+

Parallel Planning:A New Motion Planning Framework for Autonomous Driving 被引量:9

Parallel Planning:A New Motion Planning Framework for Autonomous Driving
下载PDF
导出
摘要 Motion planning is one of the most significant technologies for autonomous driving. To make motion planning models able to learn from the environment and to deal with emergency situations, a new motion planning framework called as"parallel planning" is proposed in this paper. In order to generate sufficient and various training samples, artificial traffic scenes are firstly constructed based on the knowledge from the reality.A deep planning model which combines a convolutional neural network(CNN) with the Long Short-Term Memory module(LSTM) is developed to make planning decisions in an end-toend mode. This model can learn from both real and artificial traffic scenes and imitate the driving style of human drivers.Moreover, a parallel deep reinforcement learning approach is also presented to improve the robustness of planning model and reduce the error rate. To handle emergency situations, a hybrid generative model including a variational auto-encoder(VAE) and a generative adversarial network(GAN) is utilized to learn from virtual emergencies generated in artificial traffic scenes. While an autonomous vehicle is moving, the hybrid generative model generates multiple video clips in parallel, which correspond to different potential emergency scenarios. Simultaneously, the deep planning model makes planning decisions for both virtual and current real scenes. The final planning decision is determined by analysis of real observations. Leveraging the parallel planning approach, the planner is able to make rational decisions without heavy calculation burden when an emergency occurs. Motion planning is one of the most significant technologies for autonomous driving. To make motion planning models able to learn from the environment and to deal with emergency situations, a new motion planning framework called as"parallel planning" is proposed in this paper. In order to generate sufficient and various training samples, artificial traffic scenes are firstly constructed based on the knowledge from the reality.A deep planning model which combines a convolutional neural network(CNN) with the Long Short-Term Memory module(LSTM) is developed to make planning decisions in an end-toend mode. This model can learn from both real and artificial traffic scenes and imitate the driving style of human drivers.Moreover, a parallel deep reinforcement learning approach is also presented to improve the robustness of planning model and reduce the error rate. To handle emergency situations, a hybrid generative model including a variational auto-encoder(VAE) and a generative adversarial network(GAN) is utilized to learn from virtual emergencies generated in artificial traffic scenes. While an autonomous vehicle is moving, the hybrid generative model generates multiple video clips in parallel, which correspond to different potential emergency scenarios. Simultaneously, the deep planning model makes planning decisions for both virtual and current real scenes. The final planning decision is determined by analysis of real observations. Leveraging the parallel planning approach, the planner is able to make rational decisions without heavy calculation burden when an emergency occurs.
出处 《IEEE/CAA Journal of Automatica Sinica》 EI CSCD 2019年第1期236-246,共11页 自动化学报(英文版)
基金 supported in part by the National Natural Science Foundation of China (61773414,61806076) Hubei Provincial Natural Science Foundation of China (2018CFB158)
关键词 AUTONOMOUS driving artificial traffic SCENE deep learning EMERGENCIES motion PLANNING PARALLEL PLANNING Autonomous driving artificial traffic scene deep learning emergencies motion planning parallel planning
  • 相关文献

同被引文献200

引证文献9

二级引证文献193

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部