Matching trajectories of anatomical landmarks under viewpoint, anthropometric and temporal transforms
Abstract
An approach is presented to match imaged trajectories of anatomical landmarks (e.g. hands, shoulders and feet) using semantic correspondences between human bodies. These correspondences are used to provide geometric constraints for matching actions observed from different viewpoints and performed at different rates by actors of differing anthropometric proportions. The fact that the human body has approximate anthropometric proportion allows innovative use of the machinery of epipolar geometry to provide constraints for analyzing actions performed by people of different sizes, while ensuring that changes in viewpoint do not affect matching. In addition, for linear time warps, a novel measure, constructed only from image measurements of the locations of anatomical landmarks across time, is proposed to ensure that similar actions performed at different rates are accurately matched as well. An additional feature of this new measure is that two actions from cameras moving at constant (and possibly different) velocities can also be matched. Finally, we describe how dynamic time warping can be used in conjunction with the proposed measure to match actions in the presence of nonlinear time warps. We demonstrate the versatility of our algorithm in a number of challenging sequences and applications, and report quantitative evaluation of the matching approach presented.
BibTeX
@article{Gritai-2009-122247,author = {Alexei Gritai and Yaser Sheikh and Cen Rao and Mubarak Shah},
title = {Matching trajectories of anatomical landmarks under viewpoint, anthropometric and temporal transforms},
journal = {International Journal of Computer Vision},
year = {2009},
month = {September},
volume = {84},
number = {3},
pages = {325 - 343},
}