Modeling Camera Effects to Improve Deep Vision for Real and Synthetic Data - Robotics Institute Carnegie Mellon University

Modeling Camera Effects to Improve Deep Vision for Real and Synthetic Data

Alexandra Carlson, Katherine A. Skinner, Ram Vasudevan, and M. Johnson-Roberson
Workshop Paper, ECCV '18 Workshop on Visual Learning and Embodied Agents in Simulation Environments, pp. 505 - 520, September, 2018

Abstract

Recent work has focused on generating synthetic imagery to increase the size and variability of training data for learning visual tasks in urban scenes. This includes increasing the occurrence of occlusions or varying environmental and weather effects. However, few have addressed modeling variation in the sensor domain. Sensor effects can degrade real images, limiting generalizability of network performance on visual tasks trained on synthetic data and tested in real environments. This paper proposes an efficient, automatic, physically-based augmentation pipeline to vary sensor effects – chromatic aberration, blur, exposure, noise, and color temperature – for synthetic imagery. In particular, this paper illustrates that augmenting synthetic training datasets with the proposed pipeline reduces the domain gap between synthetic and real domains for the task of object detection in urban driving scenes.

BibTeX

@workshop{Carlson-2018-130156,
author = {Alexandra Carlson and Katherine A. Skinner and Ram Vasudevan and M. Johnson-Roberson},
title = {Modeling Camera Effects to Improve Deep Vision for Real and Synthetic Data},
booktitle = {Proceedings of ECCV '18 Workshop on Visual Learning and Embodied Agents in Simulation Environments},
year = {2018},
month = {September},
pages = {505 - 520},
}