Beyond Categories: The Visual Memex Model for Reasoning About Object Relationships - Robotics Institute Carnegie Mellon University

Beyond Categories: The Visual Memex Model for Reasoning About Object Relationships

Tomasz Malisiewicz and Alexei A. Efros
Conference Paper, Proceedings of (NeurIPS) Neural Information Processing Systems, pp. 1222 - 1230, December, 2009

Abstract

The use of context is critical for scene understanding in computer vision, where the recognition of an object is driven by both local appearance and the object's relationship to other elements of the scene (context). Most current approaches rely on modeling the relationships between object categories as a source of context. In this paper we seek to move beyond categories to provide a richer appearance-based model of context. We present an exemplar-based model of objects and their relationships, the Visual Memex, that encodes both local appearance and 2D spatial context between object instances. We evaluate our model on Torralba's proposed Context Challenge against a baseline category-based system. Our experiments suggest that moving beyond categories for context modeling appears to be quite beneficial, and may be the critical missing ingredient in scene understanding systems.

BibTeX

@conference{Malisiewicz-2009-10372,
author = {Tomasz Malisiewicz and Alexei A. Efros},
title = {Beyond Categories: The Visual Memex Model for Reasoning About Object Relationships},
booktitle = {Proceedings of (NeurIPS) Neural Information Processing Systems},
year = {2009},
month = {December},
pages = {1222 - 1230},
keywords = {computer vision, object recognition, scene understanding, object context},
}