🤖 AI Summary
Existing urban-scene semantic segmentation methods primarily focus on images or point clouds, while textured meshes—rich in both geometric and appearance information—remain underexplored, especially lacking part-level annotated benchmarks. To address this gap, we introduce SUM Parts, the first large-scale, part-level semantic segmentation benchmark for urban textured meshes, covering 2.5 km² and featuring 21 fine-grained part categories with dual annotations at both mesh-patch and texture-map levels. We establish the first systematic evaluation framework tailored to part-level segmentation on textured meshes, including an in-house interactive tool for joint patch-and-texture annotation, a unified pipeline for mesh-patch labeling, texture-mapping alignment, and 3D segmentation model evaluation. Both the dataset and benchmark are publicly released, significantly advancing mesh-native semantic understanding. We conduct comprehensive evaluations across diverse 3D segmentation and interactive annotation methods, providing rigorous performance analysis and establishing new baselines.
📝 Abstract
Semantic segmentation in urban scene analysis has mainly focused on images or point clouds, while textured meshes - offering richer spatial representation - remain underexplored. This paper introduces SUM Parts, the first large-scale dataset for urban textured meshes with part-level semantic labels, covering about 2.5 km2 with 21 classes. The dataset was created using our own annotation tool, which supports both face- and texture-based annotations with efficient interactive selection. We also provide a comprehensive evaluation of 3D semantic segmentation and interactive annotation methods on this dataset. Our project page is available at https://tudelft3d.github.io/SUMParts/.