International Conference on 3D Vision (3DCV) 2025
Technical University of Munich
We introduce TrueCity, the first urban semantic segmentation benchmark with cm-accurate annotated real-world point clouds, semantic 3D city models, and annotated simulated point clouds representing the same city. TrueCity proposes segmentation classes aligned with international 3D city modeling standards, enabling consistent evaluation of synthetic-to-real gap. Our extensive experiments on common baselines quantify domain shift and highlight strategies for exploiting synthetic data to enhance real-world 3D scene understanding.
Real-world point cloud (2nd row), which was manually labeled according to the class list, used for manual modeling of semantic 3D models (3rd row), which in turn were used to simulate and auto-label synthetic point clouds (4th row).
Top-down schematic of S--R mixtures along a continuous streetscape. Solid lines mark train/validation/test splits; dashed lines mark boundaries between contiguous synthetic and real segments for each mixture ratio.
We evaluate on TrueCity under controlled synthetic-real (S--R) mixtures: 100S--0R, 75S--25R, 50S--50R, 25S--75R, and 0S--100R. We test on point-based (PointNet, PointNet++, RandLA-Net), kernel-based (KPConv), and transformer-based (Point Transformer v1/v3, Superpoint Transformer, OctFormer) methods.
Qualitative impact of the synthetic–real (S--R) training mix on models from different methods (Point-based, Kernel-based and Transformer-based).
If you use TrueCity in your research, please cite our work:
This work was developed by the TUM Chair of Geoinformatics. We thank all contributors and reviewers for their valuable feedback.