Abstract
In this paper, we present a novel unsupervised learning method for pixelization. Due to the difficulty in creating pixel art, preparing the paired training data for supervised learning is impractical. Instead, we propose an unsupervised learning framework to circumvent such difficulty. We leverage the dual nature of the pixelization and depixelization, and model these two tasks in the same network in a bi-directional manner with the input itself as training supervision. These two tasks are modeled as a cascaded network which consists of three stages for different purposes. GridNet transfers the input image into multi-scale grid-structured images with different aliasing effects. PixelNet associated with GridNet to synthesize pixel arts with sharp edges and perceptually optimal local structures. DepixelNet connects the previous network and aims to recover the pixelized result to the original image. For the sake of unsupervised learning, the mirror loss is proposed to hold the reversibility of feature representations in the process. In addition, adversarial, L1, and gradient losses are involved in the network to obtain pixel arts by retaining color correctness and smoothness. We show that our technique can synthesize crisper and perceptually more appropriate pixel arts than state-of-the-art image downscaling methods. We evaluate the proposed method with extensive experiments on many images. The proposed method outperforms state-of-the-art methods in terms of visual quality and user preference.
Original language | English |
---|---|
Number of pages | 11 |
Journal | ACM Transactions on Graphics |
Volume | 37 |
Issue number | 6 |
DOIs | |
Publication status | Published - 4 Dec 2018 |
Externally published | Yes |
Event | ACM SIGGRAPH Conference and Exhibition on Computer Graphics and Interactive Techniques in Asia 2018 - Tokyo, Japan Duration: 4 Jul 2018 → 7 Jul 2018 Conference number: 11th https://sa2018.siggraph.org/ |
Keywords
- Generative adversarial network
- Image-to-image translation
- Pixelization