Building Modeling through Enclosure Reasoning - Robotics Institute Carnegie Mellon University

Building Modeling through Enclosure Reasoning

Workshop Paper, 3DV '14 Workshop on 3D Computer Vision in the Built Environment, October, 2014

Abstract

This paper introduces a method for automatically transforming a point cloud from a laser scanner into a volumetric 3D building model based on the new concept of enclosure reasoning. Rather than simply classifying and modeling building surfaces independently or with pairwise contextual relationships, this work introduces room, floor and building level reasoning. Enclosure reasoning premises that rooms are cycles of walls enclosing free interior space. These cycles should be of minimum description length (MDL) and obey the statistical priors expected for rooms. Floors and buildings then contain the best coverage of the mostly likely rooms. This allows the pipeline to generate higher fidelity models by performing modeling and recognition jointly over the entire building at once. The complete pipeline takes raw, registered laser scan surveys of a single building. It extracts the most likely smooth architectural surfaces, locates the building, and generates wall hypotheses. The algorithm then optimizes the model by growing, merging, and pruning these hypotheses to generate the most likely rooms, floors, and building in the presence of significant clutter.

BibTeX

@workshop{Stambler-2014-7946,
author = {Adam Stambler and Daniel Huber},
title = {Building Modeling through Enclosure Reasoning},
booktitle = {Proceedings of 3DV '14 Workshop on 3D Computer Vision in the Built Environment},
year = {2014},
month = {October},
}