arXiv Analytics

Sign in

arXiv:2303.16280 [cs.CV]AbstractReferencesReviewsResources

Rethinking CycleGAN: Improving Quality of GANs for Unpaired Image-to-Image Translation

Dmitrii Torbunov, Yi Huang, Huan-Hsin Tseng, Haiwang Yu, Jin Huang, Shinjae Yoo, Meifeng Lin, Brett Viren, Yihui Ren

Published 2023-03-28Version 1

An unpaired image-to-image (I2I) translation technique seeks to find a mapping between two domains of data in a fully unsupervised manner. While the initial solutions to the I2I problem were provided by the generative adversarial neural networks (GANs), currently, diffusion models (DM) hold the state-of-the-art status on the I2I translation benchmarks in terms of FID. Yet, they suffer from some limitations, such as not using data from the source domain during the training, or maintaining consistency of the source and translated images only via simple pixel-wise errors. This work revisits the classic CycleGAN model and equips it with recent advancements in model architectures and model training procedures. The revised model is shown to significantly outperform other advanced GAN- and DM-based competitors on a variety of benchmarks. In the case of Male2Female translation of CelebA, the model achieves over 40% improvement in FID score compared to the state-of-the-art results. This work also demonstrates the ineffectiveness of the pixel-wise I2I translation faithfulness metrics and suggests their revision. The code and trained models are available at https://github.com/LS4GAN/uvcgan2

Related articles: Most relevant | Search more
arXiv:1903.04294 [cs.CV] (Published 2019-03-08)
Mix and match networks: multi-domain alignment for unpaired image-to-image translation
arXiv:2301.04685 [cs.CV] (Published 2023-01-11)
SHUNIT: Style Harmonization for Unpaired Image-to-Image Translation
arXiv:2003.04858 [cs.CV] (Published 2020-03-10)
Unpaired Image-to-Image Translation using Adversarial Consistency Loss