Bridging Scene Generation and Planning:
Driving with World Model via Unifying Vision and Motion Representation
Xingtai Gui1, Meijie Zhang2, Tianyi Yan1, Wencheng Han1, Jiahao Gong2, Feiyang Tan2, Cheng-zhong Xu1, Jianbing Shen1
1SKL-IOTSC, CIS, University of Macau, 2Afari Intelligent Drive
[2026.3.17] Release the Arxiv Ppaer
[2026.3.15] Release the WorldDrive Evaluation and Visualization script
[2026.3.14] Release the WorldDrive Project!
- News
- Table of Contents
- Abstract
- Overview
- Getting Started
- Checkpoint
- Quick Evaluation
- Visualize WorldDrive
- Quick Training
- Contact
- Acknowledgement
- Citation
End-to-end autonomous driving aims to generate safe and plausible planning policies from raw sensor input, and constructing an effective scene representation is a critical challenge. Driving world models have shown great potential in learning rich representations by predicting the future evolution of a driving scene. However, existing driving world models primarily focus on visual scene representation, and motion representation is not explicitly designed to be planner-shared and inheritable, leaving a schism between the optimization of visual scene generation and the requirements of precise motion planning. We present WorldDrive, a holistic framework that couples scene generation and real-time planning via unifying vision and motion representation. We first introduce a Trajectory-aware Driving World Model, which conditions on a trajectory vocabulary to enforce consistency between visual dynamics and motion intentions, enabling the generation of diverse and plausible future scenes conditioned on a specific trajectory. We transfer the vision and motion encoders to a downstream Multi-modal Planner, ensuring the driving policy operates on mature representations pre-optimized by scene generation. A simple interaction between motion representation, visual representation, and ego status can generate high-quality, multi-modal trajectories. Furthermore, to exploit the world model’s foresight, we propose a Future-aware Rewarder, which distills future latent representation from the frozen world model to evaluate and select optimal trajectories in real-time. Extensive experiments on the NAVSIM, NAVSIM-v2, and nuScenes benchmarks demonstrate that WorldDrive achieves state-of-the-art planning performance among vision-only methods while maintaining high-fidelity action-controlled video generation capabilities, providing strong evidence for the effectiveness of unifying vision and motion representation for robust autonomous driving.
We provide detailed guides to help you quickly set up, and evaluate WorldDrive:
- Getting started from NAVSIM environment preparation
- Preparation of WorldDrive environment
- WorldDrive Training and Evaluation
# worlddrive_stage1_train.ckpt planner checkpoint
# worlddrive_stage2_train.ckpt planner with future-aware rewarder checkpoint
# worldtraj_stage1_1024_tadwm.pkl TA-DWM pretrain checkpointDownload the pretrained 3D Causal VAE from offical CogvideoX-2B HF
👉 CogvideoX-2B VAE
sh scripts/cache/run_caching_trajworld_eval.sh # navtest for eval# download worlddrive_stage1_train.ckpt
sh scripts/evaluation/run_worlddrive_planner_pdm_score_evaluation_stage1.sh# download worlddrive_stage2_train.ckpt
sh scripts/evaluation/run_worlddrive_planner_pdm_score_evaluation_stage2.shGenerate planning result and corresponding future scene
sh scripts/visualization/worlddrive_visual.shDownload the anchor and corresponding formated PDMS
👉 Anchors
sh scripts/cache/run_caching_trajworld.sh # navtrainDownload the corresponding ta-dwm checkpoint training on NAVSIM (worldtraj_stage1_1024_tadwm) or use the checkpoint training from ta-dwm training.
👉 TA-DWM Model
sh scripts/training/run_worlddrive_planner.shIf you have any questions, please contact Xingtai via email ([email protected])
We thank the research community for their valuable support. WorldDrive is built upon the following outstanding open-source projects:
diffusers
WoTE(End-to-End Driving with Online Trajectory Evaluation via BEV World Model (ICCV2025))
Epona(Epona: Autoregressive Diffusion World Model for Autonomous Driving)
Recogdrive(A Reinforced Cognitive Framework for End-to-End Autonomous Driving) \
If you find WorldDrive is useful in your research or applications, please consider giving us a star 🌟.
