3D Hand Pose Detection in Egocentric RGB-D Images - Robotics Institute Carnegie Mellon University

3D Hand Pose Detection in Egocentric RGB-D Images

Gregory Rogez, James S. Supancic, Maryam Khademi, Jose Maria Martinez Montiel, and Deva Ramanan
Workshop Paper, ECCV '14 Workshop on Consumer Depth Cameras for Computer Vision, pp. 356 - 371, September, 2014

Abstract

We focus on the task of everyday hand pose estimation from egocentric viewpoints. For this task, we show that depth sensors are particularly informative for extracting near-field interactions of the camera wearer with his/her environment. Despite the recent advances in full-body pose estimation using Kinect-like sensors, reliable monocular hand pose estimation in RGB-D images is still an unsolved problem. The problem is considerably exacerbated when analyzing hands performing daily activities from a first-person viewpoint, due to severe occlusions arising from object manipulations and a limited field-of-view. Our system addresses these difficulties by exploiting strong priors over viewpoint and pose in a discriminative tracking-by-detection framework. Our priors are operationalized through a photorealistic synthetic model of egocentric scenes, which is used to generate training data for learning depth-based pose classifiers. We evaluate our approach on an annotated dataset of real egocentric object manipulation scenes and compare to both commercial and academic approaches. Our method provides state-of-the-art performance for both hand detection and pose estimation in egocentric RGB-D images.

BibTeX

@workshop{Rogez-2014-121190,
author = {Gregory Rogez and James S. Supancic and Maryam Khademi and Jose Maria Martinez Montiel and Deva Ramanan},
title = {3D Hand Pose Detection in Egocentric RGB-D Images},
booktitle = {Proceedings of ECCV '14 Workshop on Consumer Depth Cameras for Computer Vision},
year = {2014},
month = {September},
pages = {356 - 371},
}