An Ego-vision System for Hand Grasp Analysis
Abstract
This paper presents an egocentric vision (ego-vision) system for hand grasp analysis in unstructured environments. Our goal is to automatically recognize hand grasp types and to discover the visual structures of hand grasps using a wearable camera. In the proposed system, free hand-object interactions are recorded from a first-person viewing perspective. State-of-the-art computer vision techniques are used to detect hands and extract hand-based features. A new feature representation that incorporates hand tracking information is also proposed. Then, grasp classifiers are trained to discriminate among different grasp types from a predefined grasp taxonomy. Based on the trained grasp classifiers, visual structures of hand grasps are learned using an iterative grasp clustering method. In experiments, grasp recognition performance in both laboratory and real-world scenarios is evaluated. The best classification accuracy our system achieves is 92% and 59%, respectively. System generality to different tasks and users is also verified by the experiments. Analysis in a real-world scenario shows that it is possible to automatically learn intuitive visual grasp structures that are consistent with expert-designed grasp taxonomies.
BibTeX
@article{Cai-2017-109767,author = {Minjie Cai and Kris M. Kitani and Yoichi Sato},
title = {An Ego-vision System for Hand Grasp Analysis},
journal = {IEEE Transactions on Human-Machine Systems},
year = {2017},
month = {August},
volume = {47},
number = {4},
pages = {524 - 535},
}