TrueCity: Real and Simulated Urban Data for Cross-Domain 3D Scene Understanding

International Conference on 3D Vision (3DCV) 2025

Duc Nguyen, Yan-Ling Lai, Qilin Zhang, Prabin Gyawali, Fan Wang, Benedikt Schwab, Olaf Wysocki, Thomas H. Kolbe

Technical University of Munich

TrueCity Teaser

TrueCity introduces real-world annotated point clouds, a semantic 3D city model, and 3D-model simulated point clouds of the same location, enabling coherent evaluation of the sim-to-real domain gap in 3D scene understanding.

Abstract

We introduce TrueCity, the first urban semantic segmentation benchmark with cm-accurate annotated real-world point clouds, semantic 3D city models, and annotated simulated point clouds representing the same city. TrueCity proposes segmentation classes aligned with international 3D city modeling standards, enabling consistent evaluation of synthetic-to-real gap. Our extensive experiments on common baselines quantify domain shift and highlight strategies for exploiting synthetic data to enhance real-world 3D scene understanding.

TrueCity Benchmark Dataset

Benchmark Overview

Real-world point cloud (2nd row), which was manually labeled according to the class list, used for manual modeling of semantic 3D models (3rd row), which in turn were used to simulate and auto-label synthetic point clouds (4th row).

Key Features

Experiments

Data Split

Top-down schematic of S--R mixtures along a continuous streetscape. Solid lines mark train/validation/test splits; dashed lines mark boundaries between contiguous synthetic and real segments for each mixture ratio.

Evaluation Protocol

We evaluate on TrueCity under controlled synthetic-real (S--R) mixtures: 100S--0R, 75S--25R, 50S--50R, 25S--75R, and 0S--100R. We test on point-based (PointNet, PointNet++, RandLA-Net), kernel-based (KPConv), and transformer-based (Point Transformer v1/v3, Superpoint Transformer, OctFormer) methods.

Results

Experimental Results

Qualitative impact of the synthetic–real (S--R) training mix on models from different methods (Point-based, Kernel-based and Transformer-based).

Key Findings

Citation

If you use TrueCity in your research, please cite our work:

@inproceedings{truecity2025,
 title={TrueCity: Real and Simulated Urban Data for Cross-Domain 3D Scene Understanding},
 author={Duc Nguyen and Yan-Ling Lai and Qilin Zhang and Prabin Gyawali and Benedikt Schwab and Olaf Wysocki and Thomas H. Kolbe},
 booktitle={International Conference on 3D Vision (3DCV)},
 year={2025},
 url={https://arxiv.org/abs/2511.07007}
}

Acknowledgments

This work was developed by the TUM Chair of Geoinformatics. We thank all contributors and reviewers for their valuable feedback.