Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Interpretable and Flexible Target-Conditioned Neural Planners For Autonomous Vehicles (2309.13485v1)

Published 23 Sep 2023 in cs.RO and cs.LG

Abstract: Learning-based approaches to autonomous vehicle planners have the potential to scale to many complicated real-world driving scenarios by leveraging huge amounts of driver demonstrations. However, prior work only learns to estimate a single planning trajectory, while there may be multiple acceptable plans in real-world scenarios. To solve the problem, we propose an interpretable neural planner to regress a heatmap, which effectively represents multiple potential goals in the bird's-eye view of an autonomous vehicle. The planner employs an adaptive Gaussian kernel and relaxed hourglass loss to better capture the uncertainty of planning problems. We also use a negative Gaussian kernel to add supervision to the heatmap regression, enabling the model to learn collision avoidance effectively. Our systematic evaluation on the Lyft Open Dataset across a diverse range of real-world driving scenarios shows that our model achieves a safer and more flexible driving performance than prior works.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (26)
  1. Chauffeurnet: Learning to drive by imitating the best and synthesizing the worst. Robotics: Science and Systems, 2019.
  2. A systematic study of the class imbalance problem in convolutional neural networks. Neural Networks, 106:249–259, 2018.
  3. A. Bulat and G. Tzimiropoulos. How far are we from solving the 2d & 3d face alignment problem?(and a dataset of 230,000 3d facial landmarks). In Proceedings of the IEEE International Conference on Computer Vision, pages 1021–1030, 2017.
  4. Multimodal trajectory predictions for autonomous driving using deep convolutional networks. In 2019 International Conference on Robotics and Automation (ICRA), pages 2090–2096, 2019.
  5. Densetnt: End-to-end trajectory prediction from dense goal sets. In 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pages 15283–15292, 2021.
  6. Deep residual learning for image recognition. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770–778, 2016.
  7. One thousand and one hours: Self-driving motion prediction dataset. In Conference on Robot Learning, 2020.
  8. Driving behavior modeling using naturalistic human driving data with inverse reinforcement learning. IEEE Transactions on Intelligent Transportation Systems, 2021.
  9. Desire: Distant future prediction in dynamic scenes with interacting agents. 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 2165–2174, 2017.
  10. Unifying voxel-based representation with transformer for 3d object detection, 2022.
  11. Inferring the latent structure of human decision-making from raw visual inputs. ArXiv, abs/1703.08840, 2017.
  12. Bevformer: Learning bird’s-eye-view representation from multi-camera images via spatiotemporal transformers. In Computer Vision – ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part IX, page 1–18, Berlin, Heidelberg, 2022. Springer-Verlag.
  13. Rethinking the heatmap regression for bottom-up human pose estimation. 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 13259–13268, 2020.
  14. Overcoming limitations of mixture density networks: A sampling and fitting framework for multimodal future prediction. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 7137–7146, 2019.
  15. Motion planning for autonomous driving with a conformal spatiotemporal lattice. In 2011 IEEE International Conference on Robotics and Automation, pages 4889–4895, 2011.
  16. Numerical coordinate regression with convolutional neural networks. CoRR, abs/1801.07372, 2018.
  17. Bevsegformer: Bird’s eye view semantic segmentation from arbitrary camera rigs, 2022.
  18. D. Pomerleau. Alvinn: An autonomous land vehicle in a neural network. In D. Touretzky, editor, Proceedings of (NeurIPS) Neural Information Processing Systems, pages 305 – 313. Morgan Kaufmann, December 1989.
  19. Bev-modnet: Monocular camera based bird’s eye view moving object detection for autonomous driving. 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), pages 1503–1508, 2021.
  20. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pages 234–241. Springer, 2015.
  21. Jointly learnable behavior and trajectory planning for self-driving vehicles. 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 3949–3956, 2019.
  22. A reinforcement learning based approach for automated lane change maneuvers. In 2018 IEEE Intelligent Vehicles Symposium (IV), pages 1379–1384, 2018.
  23. A survey of autonomous driving: Common practices and emerging technologies. IEEE Access, 8:58443–58469, 2020.
  24. End-to-end interpretable neural motion planner. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 8652–8661, 2019.
  25. Tnt: Target-driven trajectory prediction. In Conference on Robot Learning, pages 895–904. PMLR, 2021.
  26. Exploring imitation learning for autonomous driving with feedback synthesizer and differentiable rasterization. In 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 1450–1457. IEEE, 2021.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Haolan Liu (6 papers)
  2. Jishen Zhao (24 papers)
  3. Liangjun Zhang (51 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.