Parsing videos of actions with segmental grammars
Abstract
Real-world videos of human activities exhibit temporal structure at various scales, long videos are typically composed out of multiple action instances, where each instance is itself composed of sub-actions with variable durations and orderings. Temporal grammars can presumably model such hierarchical structure, but are computationally difficult to apply for long video streams. We describe simple grammars that capture hierarchical temporal structure while admitting inference with a finite-state-machine. This makes parsing linear time, constant storage, and naturally online. We train grammar parameters using a latent structural SVM, where latent subactions are learned automatically. We illustrate the effectiveness of our approach over common baselines on a new half-million frame dataset of continuous YouTube videos.
BibTeX
@conference{Pirsiavash-2014-121191,author = {Hamed Pirsiavash and Deva Ramanan},
title = {Parsing videos of actions with segmental grammars},
booktitle = {Proceedings of (CVPR) Computer Vision and Pattern Recognition},
year = {2014},
month = {June},
pages = {612 - 619},
}