Multimodal Activity Recognition

12 papers with code • 10 benchmarks • 7 datasets

This task has no description! Would you like to contribute one?

Latest papers with no code

Uncertainty aware audiovisual activity recognition using deep Bayesian variational inference

no code yet • 27 Nov 2018

In the multimodal setting, the proposed framework improved precision-recall AUC by 10. 2% on the subset of MiT dataset as compared to non-Bayesian baseline.

Recognizing Human Actions as the Evolution of Pose Estimation Maps

no code yet • CVPR 2018

Specifically, the evolution of pose estimation maps can be decomposed as an evolution of heatmaps, e. g., probabilistic maps, and an evolution of estimated 2D human poses, which denote the changes of body shape and body pose, respectively.

Deep Multimodal Feature Analysis for Action Recognition in RGB+D Videos

no code yet • 23 Mar 2016

Single modality action recognition on RGB or depth sequences has been extensively explored recently.

Moving poselets: A discriminative and interpretable skeletal motion representation for action recognition

no code yet • 2015 IEEE International Conference on Computer Vision Workshop (ICCVW) 2015

While automatic feature learning methods such as supervised sparse dictionary learning or neural networks can be applied to learn feature representation and action classifiers jointly, the resulting features are usually uninterpretable.

Action recognition from depth maps using deep convolutional neural networks

no code yet • IEEE Transactions on Human-Machine Systems 2016 2015

In addition, the method was evaluated on the large dataset constructed from the above datasets.

Multimodal Multipart Learning for Action Recognition in Depth Videos

no code yet • 31 Jul 2015

We propose a joint sparse regression based learning method which utilizes the structured sparsity to model each action as a combination of multimodal features from a sparse set of body parts.

Fusing multiple features for depth-based action recognition

no code yet • ACM Transactions on Intelligent Systems and Technology (TIST) - Special Section on Visual Understanding with RGB-D Sensors 2015

The experiments are conducted on four challenging depth action databases, in order to evaluate and find the best fusion methods generally.

Group sparsity and geometry constrained dictionary learning for action recognition from depth maps.

no code yet • 2013 IEEE International Conference on Computer Vision 2014

Human action recognition based on the depth information provided by commodity depth sensors is an important yet challenging task.