3D Object Detection
583 papers with code • 55 benchmarks • 48 datasets
3D Object Detection is a task in computer vision where the goal is to identify and locate objects in a 3D environment based on their shape, location, and orientation. It involves detecting the presence of objects and determining their location in the 3D space in real-time. This task is crucial for applications such as autonomous vehicles, robotics, and augmented reality.
( Image credit: AVOD )
Libraries
Use these libraries to find 3D Object Detection models and implementationsSubtasks
Latest papers with no code
Leveraging 3D LiDAR Sensors to Enable Enhanced Urban Safety and Public Health: Pedestrian Monitoring and Abnormal Activity Detection
The integration of Light Detection and Ranging (LiDAR) and Internet of Things (IoT) technologies offers transformative opportunities for public health informatics in urban safety and pedestrian well-being.
VFMM3D: Releasing the Potential of Image by Vision Foundation Model for Monocular 3D Object Detection
Therefore, an effective solution involves transforming monocular images into LiDAR-like representations and employing a LiDAR-based 3D object detector to predict the 3D coordinates of objects.
Run-time Monitoring of 3D Object Detection in Automated Driving Systems Using Early Layer Neural Activation Patterns
To address the real-time operation requirements in ADS, we also introduce a novel introspection method that combines activation patterns from multiple layers of the detector's backbone and report its performance.
Sparse Points to Dense Clouds: Enhancing 3D Detection with Limited LiDAR Data
Our method requires only a small number of 3D points, that can be obtained from a low-cost, low-resolution sensor.
Label-Efficient 3D Object Detection For Road-Side Units
We address this challenge by devising a label-efficient object detection method for RSU based on unsupervised object discovery.
MOSE: Boosting Vision-based Roadside 3D Object Detection with Scene Cues
3D object detection based on roadside cameras is an additional way for autonomous driving to alleviate the challenges of occlusion and short perception range from vehicle cameras.
MonoTAKD: Teaching Assistant Knowledge Distillation for Monocular 3D Object Detection
Subsequently, we introduce the cross-modal residual distillation to transfer the 3D spatial cues.
DifFUSER: Diffusion Model for Robust Multi-Sensor Fusion in 3D Object Detection and BEV Segmentation
Diffusion models have recently gained prominence as powerful deep generative models, demonstrating unmatched performance across various domains.
Learning Temporal Cues by Predicting Objects Move for Multi-camera 3D Object Detection
To this end, we propose a model called DAP (Detection After Prediction), consisting of a two-branch network: (i) a branch responsible for forecasting the current objects' poses given past observations and (ii) another branch that detects objects based on the current and past observations.
NeRF-MAE: Masked AutoEncoders for Self-Supervised 3D Representation Learning for Neural Radiance Fields
Given the capabilities of neural fields in densely representing a 3D scene from 2D images, we ask the question: Can we scale their self-supervised pretraining, specifically using masked autoencoders, to generate effective 3D representations from posed RGB images.