no code implementations • 7 Apr 2024 • Chiara Plizzari, Shubham Goel, Toby Perrett, Jacob Chalk, Angjoo Kanazawa, Dima Damen
As humans move around, performing their daily tasks, they are able to recall where they have positioned objects in their environment, even if these objects are currently out of sight.
1 code implementation • 28 Nov 2023 • Hanyuan Wang, Majid Mirmehdi, Dima Damen, Toby Perrett
Previous one-stage action detection approaches have modelled temporal dependencies using only the visual modality.
no code implementations • ICCV 2023 • Chiara Plizzari, Toby Perrett, Barbara Caputo, Dima Damen
We propose and address a new generalisation problem: can a model trained for action recognition successfully classify actions when they are performed within a previously unseen scenario and in a previously unseen location?
1 code implementation • CVPR 2023 • Toby Perrett, Saptarshi Sinha, Tilo Burghardt, Majid Mirmehdi, Dima Damen
We demonstrate that, unlike naturally-collected video datasets and existing long-tail image benchmarks, current video benchmarks fall short on multiple long-tailed properties.
1 code implementation • 25 Oct 2022 • Hanyuan Wang, Majid Mirmehdi, Dima Damen, Toby Perrett
We obtain state-of-the-art performance on the challenging EPIC-KITCHENS-100 action detection as well as the standard THUMOS14 action detection benchmarks, and achieve improvement on the ActivityNet-1. 3 benchmark.
no code implementations • 14 Jul 2022 • Alessandro Masullo, Toby Perrett, Tilo Burghardt, Ian Craddock, Dima Damen, Majid Mirmehdi
We propose a novel approach to multimodal sensor fusion for Ambient Assisted Living (AAL) which takes advantage of learning using privileged information (LUPI).
1 code implementation • 11 Jun 2022 • Valentin Popescu, Dima Damen, Toby Perrett
In this paper, we evaluate state-of-the-art OCR methods on Egocentric data.
1 code implementation • 2 Jan 2022 • Hanyuan Wang, Dima Damen, Majid Mirmehdi, Toby Perrett
This incorporates a novel Voting Evidence Module to locate temporal boundaries, more accurately, where temporal contextual evidence is accumulated to predict frame-level probabilities of start and end action boundaries.
2 code implementations • CVPR 2021 • Toby Perrett, Alessandro Masullo, Tilo Burghardt, Majid Mirmehdi, Dima Damen
We propose a novel approach to few-shot action recognition, finding temporally-corresponding frame tuples between the query and videos in the support set.
1 code implementation • 29 Jul 2020 • Toby Perrett, Alessandro Masullo, Tilo Burghardt, Majid Mirmehdi, Dima Damen
This produces an initialisation for fine-tuning to target which is both context-agnostic and task-generalised.
7 code implementations • 23 Jun 2020 • Dima Damen, Hazel Doughty, Giovanni Maria Farinella, Antonino Furnari, Evangelos Kazakos, Jian Ma, Davide Moltisanti, Jonathan Munro, Toby Perrett, Will Price, Michael Wray
This paper introduces the pipeline to extend the largest dataset in egocentric vision, EPIC-KITCHENS.
Ranked #6 on Action Anticipation on EPIC-KITCHENS-100
2 code implementations • 29 Apr 2020 • Dima Damen, Hazel Doughty, Giovanni Maria Farinella, Sanja Fidler, Antonino Furnari, Evangelos Kazakos, Davide Moltisanti, Jonathan Munro, Toby Perrett, Will Price, Michael Wray
Our dataset features 55 hours of video consisting of 11. 5M frames, which we densely labelled for a total of 39. 6K action segments and 454. 2K object bounding boxes.
no code implementations • 3 Oct 2019 • Alessandro Masullo, Tilo Burghardt, Toby Perrett, Dima Damen, Majid Mirmehdi
We present the first fully automated Sit-to-Stand or Stand-to-Sit (StS) analysis framework for long-term monitoring of patients in free-living environments using video silhouettes.
no code implementations • CVPR 2019 • Toby Perrett, Dima Damen
Domain alignment in convolutional networks aims to learn the degree of layer-specific feature alignment beneficial to the joint learning of source and target datasets.
2 code implementations • ECCV 2018 • Dima Damen, Hazel Doughty, Giovanni Maria Farinella, Sanja Fidler, Antonino Furnari, Evangelos Kazakos, Davide Moltisanti, Jonathan Munro, Toby Perrett, Will Price, Michael Wray
First-person vision is gaining interest as it offers a unique viewpoint on people's interaction with objects, their attention, and even intention.
no code implementations • 22 Dec 2015 • Toby Perrett, Majid Mirmehdi, Eduardo Dias
Knowledge of human presence and interaction in a vehicle is of growing interest to vehicle manufacturers for design and safety purposes.