Efficient Temporal Consistency for Streaming Video Scene Analysis - Robotics Institute Carnegie Mellon University

Efficient Temporal Consistency for Streaming Video Scene Analysis

Conference Paper, Proceedings of (ICRA) International Conference on Robotics and Automation, pp. 133 - 139, May, 2013

Abstract

We address the problem of image-based scene analysis from streaming video, as would be seen from a moving platform, in order to efficiently generate spatially and temporally consistent predictions of semantic categories over time. In contrast to previous techniques which typically address this problem in batch and/or through graphical models, we demonstrate that by learning visual similarities between pixels across frames, a simple filtering algorithm is able to achieve high performance predictions in an efficient and online/causal manner. Our technique is a meta-algorithm that can be efficiently wrapped around any scene analysis technique that produces a per-pixel semantic category distribution. We validate our approach over three different scene analysis techniques on three different datasets that contain different semantic object categories. Our experiments demonstrate that our approach is very efficient in practice and substantially improves the consistency of the predictions over time.

BibTeX

@conference{Miksik-2013-7688,
author = {Ondrej Miksik and Daniel Munoz and J. Andrew (Drew) Bagnell and Martial Hebert},
title = {Efficient Temporal Consistency for Streaming Video Scene Analysis},
booktitle = {Proceedings of (ICRA) International Conference on Robotics and Automation},
year = {2013},
month = {May},
pages = {133 - 139},
}