- [6] introduces the requirement that translations be cycle-consistent; mapping image
to domain and back again to must yield an image that is close to the original. - [7] enforces weight sharing between the early layers of the generators and later layers of the discriminators.
- [8] combines these two ideas and models each image domain using a VAE-GAN.
- [9] utilizes reconstruction loss and teacher loss instead of VAE using a pretrained teacher network to ensure the encoder output lies in a meaningful subregion.
- [10] has presented results in generating translations between multiple distribution samples. However, their generator is conditioned on supervised labels.
- symbols
- distributions
- existing and unknown joint distribution
- distributions
- architecture
Assumption: there exists a single latent code
from which images in any of the four domains can be recovered
[6] J. Zhu, T. Park, P. Isola, and A. A. Efros, “Unpaired image-to-image translation using cycleconsistent adversarial networks,” CoRR, vol. abs/1703.10593, 2017.
[7] M.-Y. Liu and O. Tuzel, “Coupled generative adversarial networks,” in Advances in Neural Information Processing Systems 29 (D. D. Lee, M. Sugiyama, U. V. Luxburg, I. Guyon, and R. Garnett, eds.), pp. 469–477, Curran Associates, Inc., 2016.
[8] M. Liu, T. Breuel, and J. Kautz, “Unsupervised image-to-image translation networks,” CoRR, vol. abs/1703.00848, 2017.
[9] A. Royer, K. Bousmalis, S. Gouws, F. Bertsch, I. Moressi, F. Cole, and K. Murphy, “Xgan: Unsupervised image-to-image translation for many-to-many mappings,” arXiv preprint arXiv:1711.05139, 2017.
[10] Y. Choi, M. Choi, M. Kim, J.-W. Ha, S. Kim, and J. Choo, “Stargan: Unified generative adversarial networks for multi-domain image-to-image translation,” arXiv preprint arXiv:1711.09020, 2017.
扩展UNIT到多分布表达