Real-Time User-Guided Image Colorization with Learned Deep Priors
Abstract
We propose a deep learning approach for user-guided image colorization. The system directly maps a grayscale image, along with sparse, local user "hints" to an output colorization with a Convolutional Neural Network (CNN). Rather than using hand-defined rules, the network propagates user edits by fusing low-level cues along with high-level semantic information, learned from large-scale data. We train on a million images, with simulated user inputs. To guide the user towards efficient input selection, the system recommends likely colors based on the input image and current user inputs. The colorization is performed in a single feed-forward pass, enabling real-time use. Even with randomly simulated user inputs, we show that the proposed system helps novice users quickly create realistic colorizations, and offers large improvements in colorization quality with just a minute of use. In addition, we demonstrate that the framework can incorporate other user "hints" to the desired colorization, showing an application to color histogram transfer.
BibTeX
@article{Zhang-2017-125694,author = {Richard Zhang and Jun-Yan Zhu and Phillip Isola and Xinyang Geng and Angela S. Lin and Tianhe Yu and Alexei A. Efros},
title = {Real-Time User-Guided Image Colorization with Learned Deep Priors},
journal = {ACM Transactions on Graphics (TOG)},
year = {2017},
month = {July},
volume = {36},
number = {4},
}