Phillip Isola [email protected] (BAIR) Laboratory University of California Berkeley, Jun-Yan Zhu [email protected] (BAIR) Laboratory University of California Berkeley, Tinghui Zhou [email protected] (BAIR) Laboratory University of California Berkeley, Alexei A Efros [email protected] (BAIR) Laboratory University of California Berkeley, Berkeley Ai Research (BAIR) Laboratory University of California Berkeley (2016)
This paper investigates image-to-image translation using Conditional Adversarial Networks (cGANs). The authors highlight the limitations of traditional Convolutional Neural Networks (CNNs) in generating sharp images when minimizing Euclidean loss and propose a framework leveraging GANs to automatically learn appropriate loss functions that encourage realism in outputs. They discuss their contributions in demonstrating the effectiveness of cGANs across various tasks and offer a simplified framework while exploring architectural choices. The method's robustness is validated through experiments on different datasets including Cityscapes and ImageNet, showing promising results for generating realistic images from various types of input images like semantic labels and sketches. They also emphasize the need for perceptual evaluation of generated images and report qualitative and quantitative results from their experiments, concluding that cGANs provide a versatile solution for many image-to-image translation tasks.
This paper employs the following methods:
The following datasets were used in this research:
The authors identified the following limitations: