First Person Action Recognition Using Deep Learned Descriptors
Citations Over TimeTop 1% of 2016 papers
Abstract
We focus on the problem of wearer's action recognition in first person a.k.a. egocentric videos. This problem is more challenging than third person activity recognition due to unavailability of wearer's pose and sharp movements in the videos caused by the natural head motion of the wearer. Carefully crafted features based on hands and objects cues for the problem have been shown to be successful for limited targeted datasets. We propose convolutional neural networks (CNNs) for end to end learning and classification of wearer's actions. The proposed network makes use of egocentric cues by capturing hand pose, head motion and saliency map. It is compact. It can also be trained from relatively small number of labeled egocentric videos that are available. We show that the proposed network can generalize and give state of the art performance on various disparate egocentric action datasets.
Related Papers
- → INCORPORATION OF RECOVERY AND REPAIR TIME IN THE RELIABILITY MODELLING OF FAULT-TOLERANT SYSTEM(1983)3 cited
- → An analytical approach to the unavailability estimation of protected connections in optical networks(2008)
- → Incorporation of Recovery and Repair Time in the Reliability Modelling of Fault-Tolerant System(1983)
- Processing mechanism for Web services unavailability based on reflective technology(2007)
- → Go in the Opposite Direction? The Impact of Unavailability on Crowdfunding Success(2020)