Pose Estimation
1334 papers with code • 28 benchmarks • 113 datasets
Pose Estimation is a computer vision task where the goal is to detect the position and orientation of a person or an object. Usually, this is done by predicting the location of specific keypoints like hands, head, elbows, etc. in case of Human Pose Estimation.
A common benchmark for this task is MPII Human Pose
( Image credit: Real-time 2D Multi-Person Pose Estimation on CPU: Lightweight OpenPose )
Libraries
Use these libraries to find Pose Estimation models and implementationsSubtasks
- 3D Human Pose Estimation
- Keypoint Detection
- 3D Pose Estimation
- 6D Pose Estimation
- 6D Pose Estimation
- Hand Pose Estimation
- 6D Pose Estimation using RGB
- Multi-Person Pose Estimation
- Head Pose Estimation
- Human Pose Forecasting
- 6D Pose Estimation using RGBD
- Animal Pose Estimation
- Vehicle Pose Estimation
- RF-based Pose Estimation
- Car Pose Estimation
- Hand Joint Reconstruction
- Activeness Detection
- Semi-supervised 2D and 3D landmark labeling
Latest papers with no code
GLID: Pre-training a Generalist Encoder-Decoder Vision Model
This paper proposes a GeneraLIst encoder-Decoder (GLID) pre-training method for better handling various downstream computer vision tasks.
Measuring proximity to standard planes during fetal brain ultrasound scanning
This paper introduces a novel pipeline designed to bring ultrasound (US) plane pose estimation closer to clinical use for more effective navigation to the standard planes (SPs) in the fetal brain.
Incremental Joint Learning of Depth, Pose and Implicit Scene Representation on Monocular Camera in Large-scale Scenes
For pose estimation, a feature-metric bundle adjustment (FBA) method is designed for accurate and robust camera tracking in large-scale scenes.
Learning 3D-Aware GANs from Unposed Images with Template Feature Field
Collecting accurate camera poses of training images has been shown to well serve the learning of 3D-aware generative adversarial networks (GANs) yet can be quite expensive in practice.
Learning a Category-level Object Pose Estimator without Pose Annotations
Instead of using manually annotated images, we leverage diffusion models (e. g., Zero-1-to-3) to generate a set of images under controlled pose differences and propose to learn our object pose estimator with those images.
Two Hands Are Better Than One: Resolving Hand to Hand Intersections via Occupancy Networks
This work addresses the intersection of hands by exploiting an occupancy network that represents the hand's volume as a continuous manifold.
Multi Positive Contrastive Learning with Pose-Consistent Generated Images
Model pre-training has become essential in various recognition tasks.
3D Congealing: 3D-Aware Image Alignment in the Wild
The framework optimizes for the canonical representation together with the pose for each input image, and a per-image coordinate map that warps 2D pixel coordinates to the 3D canonical frame to account for the shape matching.
Marrying NeRF with Feature Matching for One-step Pose Estimation
Given the image collection of an object, we aim at building a real-time image-based pose estimation method, which requires neither its CAD model nor hours of object-specific training.
OmniLocalRF: Omnidirectional Local Radiance Fields from Dynamic Videos
Omnidirectional cameras are extensively used in various applications to provide a wide field of vision.