Sun, Q., Wang, H., Zhan, J., Nie, F., Wen, X., Xu, L., Zhan, K., Jia, P., Lang, X., Zhao, H. (2024). Generalizing Motion Planners with Mixture of Experts for Autonomous Driving. arXiv preprint arXiv:2410.15774v1.
This paper investigates the generalization capabilities of learning-based motion planners for autonomous driving and aims to improve their performance in complex, few-shot, and zero-shot driving scenarios by leveraging large-scale datasets and a Mixture-of-Experts (MoE) architecture.
The researchers propose StateTransformer-2 (STR2), a scalable, decoder-only motion planner that utilizes a Vision Transformer (ViT) encoder and a MoE causal Transformer architecture. They train and evaluate STR2 on the NuPlan dataset, a large-scale dataset for autonomous driving, and benchmark its performance against several state-of-the-art motion planners. Additionally, they conduct scaling experiments on an industrial-level dataset from LiAuto, comprising billions of real-world urban driving scenarios.
The study demonstrates that scaling learning-based motion planners with MoE architectures and massive datasets significantly enhances their generalization capabilities in autonomous driving. This approach enables the development of more robust and reliable motion planners capable of handling the complexities of real-world driving environments.
This research contributes to the advancement of autonomous driving technology by presenting a scalable and generalizable motion planning approach. The findings have significant implications for developing safer and more efficient self-driving systems.
Future work includes comprehensive scaling analysis with larger models on the LiAuto dataset, exploring more advanced simulation environments for interaction-intensive scenarios, and optimizing inference time for real-time applications.
翻譯成其他語言
從原文內容
arxiv.org
深入探究