Detecting cars in aerial photographs with a hierarchy of deconvolution nets
Abstract
Detecting cars in large aerial photographs can be quite a challenging task, given that cars in such datasets are often barely visible to the naked human eye. Traditional object detection algorithms fail to perform well when it comes to detecting cars under such circumstances. One would rather use context or exploit spatial relationship between different entities in the scene to narrow down the search space. We aim to do so by looking at different resolutions of the image to process context and focus on promising areas. This is done using a hierarchy of deconvolution networks with each level of the hierarchy trying to predict a heatmap of a certain resolution. We show that our architecture is able to model context implicitly and use it for finer prediction and faster search.
BibTeX
@techreport{Chakraborty-2016-5623,author = {Satyaki Chakraborty and Daniel Maturana and Sebastian Scherer},
title = {Detecting cars in aerial photographs with a hierarchy of deconvolution nets},
year = {2016},
month = {November},
institute = {Carnegie Mellon University},
address = {Pittsburgh, PA},
number = {CMU-RI-TR-16-60},
}