3D Semantic Segmentation
169 papers with code • 14 benchmarks • 31 datasets
3D Semantic Segmentation is a computer vision task that involves dividing a 3D point cloud or 3D mesh into semantically meaningful parts or regions. The goal of 3D semantic segmentation is to identify and label different objects and parts within a 3D scene, which can be used for applications such as robotics, autonomous driving, and augmented reality.
Libraries
Use these libraries to find 3D Semantic Segmentation models and implementationsDatasets
Subtasks
Latest papers with no code
Language-Assisted 3D Scene Understanding
The scale and quality of point cloud datasets constrain the advancement of point cloud learning.
Transferring CLIP's Knowledge into Zero-Shot Point Cloud Semantic Segmentation
In this work, we focus on zero-shot point cloud semantic segmentation and propose a simple yet effective baseline to transfer the visual-linguistic knowledge implied in CLIP to point cloud encoder at both feature and output levels.
Novel class discovery meets foundation models for 3D semantic segmentation
Firstly, it introduces the novel task of NCD for point cloud semantic segmentation.
ALSTER: A Local Spatio-Temporal Expert for Online 3D Semantic Reconstruction
Using these main contributions, our method can enable scenarios with real-time constraints and can scale to arbitrary scene sizes by processing and updating the scene only in a local region defined by the new measurement.
2D Feature Distillation for Weakly- and Semi-Supervised 3D Semantic Segmentation
As 3D perception problems grow in popularity and the need for large-scale labeled datasets for LiDAR semantic segmentation increase, new methods arise that aim to reduce the necessity for dense annotations by employing weakly-supervised training.
Seeing Beyond Cancer: Multi-Institutional Validation of Object Localization and 3D Semantic Segmentation using Deep Learning for Breast MRI
The clinical management of breast cancer depends on an accurate understanding of the tumor and its anatomical context to adjacent tissues and landmark structures.
Instance-aware 3D Semantic Segmentation powered by Shape Generators and Classifiers
In the experiments, our method significantly outperform existing approaches in 3D semantic segmentation on several public benchmarks, such as Waymo Open Dataset, SemanticKITTI and ScanNetV2.
Leveraging Large-Scale Pretrained Vision Foundation Models for Label-Efficient 3D Point Cloud Segmentation
Recently, large-scale pre-trained models such as Segment-Anything Model (SAM) and Contrastive Language-Image Pre-training (CLIP) have demonstrated remarkable success and revolutionized the field of computer vision.
Revisiting Multi-modal 3D Semantic Segmentation in Real-world Autonomous Driving
LiDAR and camera are two critical sensors for multi-modal 3D semantic segmentation and are supposed to be fused efficiently and robustly to promise safety in various real-world scenarios.
Geometry Aware Field-to-field Transformations for 3D Semantic Segmentation
We present a novel approach to perform 3D semantic segmentation solely from 2D supervision by leveraging Neural Radiance Fields (NeRFs).