Skip to main content
Article
Maximum Spatial Perturbation Consistency for Unpaired Image-to-Image Translation
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
  • Yanwu Xu, University of Pittsburgh
  • Shaoan Xie, Carnegie Mellon University
  • Wenhao Wu, Baidu, Inc.
  • Kun Zhang, Carnegie Mellon University & Mohamed bin Zayed University of Artificial Intelligence
  • Mingming Gong, School of Mathematics and Statistics
  • Kayhan Batmanghelich, University of Pittsburgh
Document Type
Conference Proceeding
Abstract

Unpaired image-to-image translation (I2I) is an ill-posed problem, as an infinite number of translation functions can map the source domain distribution to the target distribution. Therefore, much effort has been put into designing suitable constraints, e.g., cycle consistency (CycleGAN), geometry consistency (GCGAN), and contrastive learning-based constraints (CUTGAN), that help better pose the problem. However, these well-known constraints have limitations: (1) they are either too restrictive or too weak for specific I2I tasks; (2) these methods result in content distortion when there is a significant spatial variation between the source and target domains. This paper proposes a universal regularization technique called maximum spatial perturbation consistency (MSPC), which enforces a spatial perturbation function (T) and the translation operator (G) to be commutative (i.e., T G=GT). In addition, we introduce two adversarial training components for learning the spatial perturbation function. The first one lets T compete with G to achieve maximum perturbation. The second one lets G and T compete with discriminators to align the spatial variations caused by the change of object size, object distortion, background interruptions, etc. Our method outperforms the state-of-the-art methods on most I2I benchmarks. We also introduce a new benchmark, namely the front face to profile face dataset, to emphasize the underlying challenges of I2I for real-world applications. We finally perform ablation experiments to study the sensitivity of our method to the severity of spatial perturbation and its effectiveness for distribution alignment.

DOI
10.1109/CVPR52688.2022.01777
Publication Date
9-27-2022
Keywords
  • Training,
  • Geometry,
  • Computer vision,
  • Sensitivity,
  • Perturbation methods,
  • Face recognition,
  • Benchmark testing
Comments

Open Access version, provided by CVPR

Archived with thanks to CVPR

License: CC by 4.0

Uploaded 27 April 2023

Citation Information
Y. Xu, S. Xie, W. Wu, K. Zhang, M. Gong and K. Batmanghelich, "Maximum Spatial Perturbation Consistency for Unpaired Image-to-Image Translation," 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 2022, pp. 18290-18299, doi: 10.1109/CVPR52688.2022.01777.