Keywords: 3D representation learning, Self-Supervised Learning
TL;DR: We learn 3D representations from procedurally generated shapes, achieving state-of-the-art performance without semantic data. Our approach underscores point cloud SSL’s strength in capturing geometry over semantics in 3D tasks.
Abstract: Self-supervised learning has emerged as a promising approach for acquiring transferable 3D representations from unlabeled 3D point clouds. Unlike 2D images, which are widely accessible, acquiring 3D assets requires specialized expertise or professional 3D scanning equipment, making it difficult to scale and raising copyright concerns. To address these challenges, we propose learning 3D representations from procedural 3D programs that automatically generate 3D shapes using simple primitives and augmentations.
Remarkably, despite lacking semantic content, the 3D representations learned from the procedurally generated 3D shapes perform on par with state-of-the-art representations learned from semantically recognizable 3D models (e.g., airplanes) across various downstream 3D tasks, including shape classification, part segmentation, and masked point cloud completion.
We provide a detailed analysis on factors that make a good 3D procedural programs.
Extensive experiments further suggest that current self-supervised learning methods on point clouds do not rely on semantics of 3D shapes, shedding light on the nature of 3D representations learned.
Submission Number: 4
Loading