3D-tracking of head and hands for pointing gesture recognition in a human-robot interaction scenario
Citations Over TimeTop 10% of 2004 papers
Abstract
We present our approach for visual tracking of head, hands and head orientation. Given the images provided by a calibrated stereo-camera, color and disparity information are integrated into a multi-hypotheses tracking framework in order to find the 3D-positions of the respective body parts. Based on the hands' motion, an HMM-based approach is applied to recognize pointing gestures. We show experimentally, that the gesture recognition performance can be improved significantly by using visually gained information about head orientation as an additional feature. Our system aims at applications in the field of human-robot interaction, where it is important to do run-on recognition in real-time, to allow for robot's egomotion and not to rely on manual initialization.
Related Papers
- → An Intelligent Smart Home Control Using Body Gestures(2006)18 cited
- → An Intelligent Smart Home Control Using Body Gestures(2006)16 cited
- → Real‐time hand gestures system based on leap motion(2018)11 cited
- → Dynamic hand gesture early recognition based on Hidden Semi-Markov Models(2014)10 cited
- Real-Time Gesture Recognition for Dynamic Applications(2015)