Object Tracking
584 papers with code • 7 benchmarks • 61 datasets
Object tracking is the task of taking an initial set of object detections, creating a unique ID for each of the initial detections, and then tracking each of the objects as they move around frames in a video, maintaining the ID assignment. State-of-the-art methods involve fusing data from RGB and event-based cameras to produce more reliable object tracking. CNN-based models using only RGB images as input are also effective. The most popular benchmark is OTB. There are several evaluation metrics specific to object tracking, including HOTA, MOTA, IDF1, and Track-mAP.
( Image credit: Towards-Realtime-MOT )
Benchmarks
These leaderboards are used to track progress in Object Tracking
Libraries
Use these libraries to find Object Tracking models and implementationsDatasets
Subtasks
Latest papers with no code
360VOTS: Visual Object Tracking and Segmentation in Omnidirectional Videos
Visual object tracking and segmentation in omnidirectional videos are challenging due to the wide field-of-view and large spherical distortion brought by 360{\deg} images.
TeamTrack: A Dataset for Multi-Sport Multi-Object Tracking in Full-pitch Videos
Multi-object tracking (MOT) is a critical and challenging task in computer vision, particularly in situations involving objects with similar appearances but diverse movements, as seen in team sports.
Inverse Neural Rendering for Explainable Multi-Object Tracking
We propose to recast 3D multi-object tracking from RGB cameras as an \emph{Inverse Rendering (IR)} problem, by optimizing via a differentiable rendering pipeline over the latent space of pre-trained 3D object representations and retrieve the latents that best represent object instances in a given input image.
MLS-Track: Multilevel Semantic Interaction in RMOT
The new trend in multi-object tracking task is to track objects of interest using natural language.
KnotResolver: Tracking self-intersecting filaments in microscopy using directed graphs
Quantification of microscopy time-series of in vitro reconstituted motor driven microtubule (MT) transport in 'gliding assays' is typically performed using computational object tracking tools.
How to deal with glare for improved perception of Autonomous Vehicles
In this paper, we investigate various glare reduction techniques, including the proposed saturated pixel-aware glare reduction technique for improved performance of the computer vision (CV) tasks employed by the perception layer of AVs.
Into the Fog: Evaluating Multiple Object Tracking Robustness
To address these limitations, we propose a pipeline for physic-based volumetric fog simulation in arbitrary real-world MOT dataset utilizing frame-by-frame monocular depth estimation and a fog formation optical model.
Gaga: Group Any Gaussians via 3D-aware Memory Bank
We introduce Gaga, a framework that reconstructs and segments open-world 3D scenes by leveraging inconsistent 2D masks predicted by zero-shot segmentation models.
Trashbusters: Deep Learning Approach for Litter Detection and Tracking
This research focuses on automating the penalization of litterbugs, addressing the persistent problem of littering in public places.
MTMMC: A Large-Scale Real-World Multi-Modal Camera Tracking Benchmark
Multi-target multi-camera tracking is a crucial task that involves identifying and tracking individuals over time using video streams from multiple cameras.