Egocentric Activity Recognition
14 papers with code • 2 benchmarks • 4 datasets
Libraries
Use these libraries to find Egocentric Activity Recognition models and implementationsLatest papers with no code
MMG-Ego4D: Multi-Modal Generalization in Egocentric Action Recognition
In this paper, we study a novel problem in egocentric action recognition, which we term as "Multimodal Generalization" (MMG).
Optical Flow Estimation in 360$^\circ$ Videos: Dataset, Model and Application
Moreover, we present a novel Siamese representation Learning framework for Omnidirectional Flow (SLOF) estimation, which is trained in a contrastive manner via a hybrid loss that combines siamese contrastive and optical flow losses.
Domain Generalization through Audio-Visual Relative Norm Alignment in First Person Action Recognition
First person action recognition is becoming an increasingly researched area thanks to the rising popularity of wearable cameras.
Egocentric Activity Recognition and Localization on a 3D Map
Given a video captured from a first person perspective and the environment context of where the video is recorded, can we recognize what the person is doing and identify where the action occurs in the 3D space?
Egok360: A 360 Egocentric Kinetic Human Activity Video Dataset
To bridge this gap, in this paper we propose a novel Egocentric (first-person) 360{\deg} Kinetic human activity video dataset (EgoK360).
Symbiotic Attention with Privileged Information for Egocentric Action Recognition
Due to the large action vocabulary in egocentric video datasets, recent studies usually utilize a two-branch structure for action recognition, ie, one branch for verb classification and the other branch for noun classification.
Self-supervising Action Recognition by Statistical Moment and Subspace Descriptors
In this paper, we build on a concept of self-supervision by taking RGB frames as input to learn to predict both action concepts and auxiliary descriptors e. g., object descriptors.
On the Role of Event Boundaries in Egocentric Activity Recognition from Photostreams
Event boundaries play a crucial role as a pre-processing step for detection, localization, and recognition tasks of human activities in videos.
Multi-modal Egocentric Activity Recognition using Audio-Visual Features
In this work, we propose a new framework for egocentric activity recognition problem based on combining audio-visual features with multi-kernel learning (MKL) and multi-kernel boosting (MKBoost).
Egocentric Activity Recognition on a Budget
Recent advances in embedded technology have enabled more pervasive machine learning.