Skip to content
/ RAD Public

[NeurIPS 2025] RAD: Training an End-to-End Driving Policy via Large-Scale 3DGS-based Reinforcement Learning

License

Notifications You must be signed in to change notification settings

hustvl/RAD

Repository files navigation

RAD: Training an End-to-End Driving Policy via Large-Scale

3DGS-based Reinforcement Learning

Hao Gao1, Shaoyu Chen1,2,†, Bo Jiang1, Bencheng Liao1, Yiang Shi1, Xiaoyang Guo2, Yuechuan Pu2, Haoran Yin2, Xiangyu Li2, Xinbang Zhang2, Ying Zhang2, Wenyu Liu1, Qian Zhang2, Xinggang Wang1,📧

1 Huazhong University of Science and Technology, 2 Horizon Robotics, Project lead 📧 Corresponding author

Project page arXiv license

📰 News

  • [2025.09.28] We have released core code for RL training.

  • [2025.09.18] RAD has been accepted by NeurIPS 2025! 🎉🎉🎉

  • [2025.02.18] We released our paper on Arxiv. Code are coming soon. Please stay tuned! ☕️

🎯 How to Use

  • Project Structure
.
├── data/                        # Action anchors for planning/control
├── compute_advantage.py         # Script for computing RL advantages and evaluation metrics
├── generate_action_anchor.py    # Script for generating action anchors for planning/control
├── planning_head.py             # Planning head module
└── README.md
  • Run Key Scripts
# You can quickly test the core functionality by running the provided scripts.
# Generate action anchors
python generate_action_anchor.py

# Run the planning head module
python planning_head.py

# Compute advantage metrics
python compute_advantage.py
  • Using Your Own Data

To integrate this project into your pipeline and use your own data, follow these steps:

  1. Replace the Planning Head
    Use planning_head.py to replace the head of your end-to-end algorithm.

  2. Prepare the Closed-Loop Environment
    Set up your closed-loop environment and collect closed-loop data.

  3. Compute Advantages and Train the Model
    Use compute_advantage.py to calculate advantage values from the collected data, and then use them for model training.

📌 RAD Training Discussion & Reference

We have created a central discussion issue for RAD training details. You can view and participate in the discussion here: RAD Training Details Issue

📚 Citation

If you find RAD useful in your research or applications, please consider giving us a star 🌟 and citing it by the following BibTeX entry.

@article{RAD,
  title={RAD: Training an End-to-End Driving Policy via Large-Scale 3DGS-based Reinforcement Learning},
  author={Gao, Hao and Chen, Shaoyu and Jiang, Bo and Liao, Bencheng and Shi, Yiang and Guo, Xiaoyang and Pu, Yuechuan and Yin, Haoran and Li, Xiangyu and Zhang, Xinbang and Zhang, Ying and Liu, Wenyu and Zhang, Qian and Wang, Xinggang},
  journal={arXiv preprint arXiv:2502.13144},
  year={2025}
}

About

[NeurIPS 2025] RAD: Training an End-to-End Driving Policy via Large-Scale 3DGS-based Reinforcement Learning

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages