Unified Driving World Model

HERMES++: Toward a Unified Driving World Model for 3D Scene Understanding and Generation

Xin Zhou1 Dingkang Liang1 Xiwu Chen2 Feiyang Tan2 Dingyuan Zhang1 Hengshuang Zhao3 Xiang Bai1
1Huazhong University of Science and Technology 2Mach Drive 3The University of Hong Kong

Abstract

Driving world models serve as a pivotal technology for autonomous driving by simulating environmental dynamics. However, existing approaches predominantly focus on future scene generation, often overlooking comprehensive 3D scene understanding. Conversely, while Large Language Models (LLMs) demonstrate impressive reasoning capabilities, they lack the capacity to predict future geometric evolution, creating a significant disparity between semantic interpretation and physical simulation. To bridge this gap, we propose HERMES++, a unified driving world model that integrates 3D scene understanding and future geometry prediction within a single framework. Our approach addresses the distinct requirements of these tasks through synergistic designs. First, a BEV representation consolidates multi-view spatial information into a structure compatible with LLMs. Second, we introduce LLM-enhanced world queries to facilitate knowledge transfer from the understanding branch. Third, a Current-to-Future Link is designed to bridge the temporal gap, conditioning geometric evolution on semantic context. Finally, to enforce structural integrity, we employ a Joint Geometric Optimization strategy that integrates explicit geometric constraints with implicit latent regularization to align internal representations with geometry-aware priors. Extensive evaluations on multiple benchmarks validate the effectiveness of our method. HERMES++ achieves strong performance, outperforming specialist approaches in both future point cloud prediction and 3D scene understanding tasks.

Unified Tasks

One model handles both 3D scene understanding and future point cloud generation.

LLM-Enhanced Queries

World queries transfer semantic knowledge from language reasoning into future geometry prediction.

Geometric Optimization

Explicit and implicit geometric constraints keep generated futures structurally consistent.

HERMES++ unified driving world model overview

Pipeline

HERMES++ pipeline

Demo

Example 1

Example 2

Main Results

HERMES++ main results

Why BEV as the LLM Input?

Compared with directly feeding multi-view image features into the LLM, the BEV representation preserves spatial structure more effectively. This improves future point cloud generation while maintaining comparable scene understanding quality.

Analysis of BEV input for HERMES++

Effect of Joint Geometric Optimization

Joint Geometric Optimization aligns latent features with geometry-aware priors. Compared with using explicit point-cloud constraints alone, the learned representations recover clearer road layouts and future structural patterns that better match LiDAR observations.

This improves the geometry branch without sacrificing the semantic reasoning ability inherited from the LLM.

Joint Geometric Optimization visualization

Qualitative Results

HERMES++ qualitative results

BibTeX

@article{zhou2026hermespp,
  title={HERMES++: Toward a Unified Driving World Model for 3D Scene Understanding and Generation},
  author={Zhou, Xin and Liang, Dingkang and Chen, Xiwu and Tan, Feiyang and Zhang, Dingyuan and Zhao, Hengshuang and Bai, Xiang},
  journal={arXiv preprint: 2604.28196},
  year={2026}
}

@inproceedings{zhou2025hermes,
  title={HERMES: A Unified Self-Driving World Model for Simultaneous 3D Scene Understanding and Generation},
  author={Zhou, Xin and Liang, Dingkang and Tu, Sifan and Chen, Xiwu and Ding, Yikang and Zhang, Dingyuan and Tan, Feiyang and Zhao, Hengshuang and Bai, Xiang},
  booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
  year={2025}
}