Wei Zhan serves as a Co-Director of Berkeley DeepDrive, one of the leading research centers in the field of AI for autonomy and mobility involving many Berkeley faculty and industrial partners. He is an Assistant Professional Researcher at UC Berkeley leading a team of Ph.D. students and Postdocs conducting research. His research is focused on AI for autonomous systems leveraging control, robotics, computer vision and machine learning techniques to tackle challenges with sophisticated dynamics, interactive human behavior and complex scenes in a scalable way. He also teaches AI for Autonomy at UC Berkeley.
He is also Chief Scientist of Applied Intuition, leading AI research efforts towards next-generation autonomy and its development toolchain. He is actively hiring Research Scientists, Research Engineers and Research Interns.
He received his Ph.D. degree from UC Berkeley. His publications received the Best Student Paper Award in IV’18, Best Paper Award – Honorable Mention of IEEE Robotics and Automation Letters, and Best Paper Award of ICRA’24. One of his publications also got ICLR’23 notable top 5% oral presentation. He led the construction of INTERACTION dataset and the organization of its prediction challenges in NeurIPS’20 and ICCV’21.
Research
Current

Policy Customization and Multi-Agent Reinforcement Learning


3D Surface Reconstruction with Unsupervised Decomposition
- DeSiRe-GS – 4D Gaussians for Decomposition and Mesh: CVPR’25, arxiv
- Q-SLAM – quadric representations for monocular SLAM: CoRL’24, RA-Letters ’23
- S3 Gaussian – Self-Supervised Street Gaussian: arxiv, Code


Interaction-Aware 3D Generation and Manipulation


Cross-Embodiment and Generalization for Robot Learning




Autonomous Racing – Learning to Plan and Control at the Limits
- Active exploration for modeling dynamics and racing behavior with Gaussian Process: IEEE Trans-CST ’24, IFAC’23, arxiv
- Skill-Critic – refining learned skills for reinforcement learning: RA-Letters ’24, Website
- BeTAIL- behavior transformer adversarial imitation learning: RA-Letters ’24, Website
Recent and Continued

3D Perception with Temporal, Multi-View, and Multi-Modal Fusion
- SOLOFusion – temporal multi-view 3D detection: ICLR’23 (notable top 5% oral presentation), arxiv, Code
- SparseFusion – fusing multi-modal sparse representations: ICCV’23, arxiv, Code
- NeRF-Det – learning geometry-aware volumetric representation: ICCV’23, arxiv, Code
- Fusing BEV point cloud and front-view image: IV’18, arxiv, Code
Efficient, Automated Data Engine and Training Pipeline

Self-Supervised Learning for Differentiable Autonomy Stack
Behavior and Scenario Generation for Closed-Loop Simulation
- Guided diffusion for traffic simulation with controllable criticality: ECCV’24, arxiv
- Editing driver character with socially-controllable generation: RA-Letters ’23, arxiv
- Diverse Critical Generation: IROS’21, arxiv, IEEE Trans-ITS ’21
Past

LiDAR-based and 2D Perception
Generalizable, Multi-Agent, Interactive Prediction
- Scenario-transferable with semantic intention representation: IEEE Tran-ITS ’22, arxiv, Video summary, IV’18 (Best Student Paper Award)
- Transferable and adaptable prediction: NeurIPS’21 ML4AD workshop (spotlight), arxiv
- Multi-agent prediction combining egocentric and allocentric views: CoRL’21
- Social posterior collapse in variational autoencoder: NeurIPS’21, arxiv
- Probabilistic prediction with hierarchical inverse reinforcement learning: ITSC’18
Decision, Planning, Control and Behavior Design
- Sample-based inverse reinforcement learning and socially compatible planner: RA-Letters ’20, IROS’18, RA-Letters ’20 (Best Paper Award – Honorable Mention)
- Constrained Iterative LQR/LQG: IEEE Trans-IV ’19, ITSC’17, IROS’21
- Imitation learning trained and combined with model predictive control: DSCC’18
- Integrated decision and planning, efficient planner: ITSC’16, ITSC’17, ITSC’17