Generating Anime Faces From Human Faces With Adversarial Networks
Generating Anime Faces From Human Faces With Adversarial Networks
Generating Anime Faces From Human Faces With Adversarial Networks
NETWORKS
1
Yu-Jing Lin (林裕景), 1 Chiou-Shann Fuh (傅楸善)
1
Department of Computer Science and Information Engineering,
National Taiwan University, Taipei, Taiwan
LD = Ex∼pdata (x) [log D(x)] (1) The critical part of unpaired domain transfer is to use a pair
of GAN. For a generated fake image in domain Y , the GY X
is supposed to have the ability of converting it back. To make
LG = Ez∼pz (z) [log (1 − D(G(z)))] (2) sure GXY and GY X converts an image to domain Y and back
The total objective function is the sum of discriminating to domain X. The cycle consistency loss is introduced as
loss and generating loss: Figure 3 shows and as the following equation:
3. FACE2ANIME
Fig. 9: Faces generated by residual block generator trained on SCUP-FBP5500. Residual block generator on SCUP-FBP5500
works as well as on CelebA, implying that it is feasible to do adversarial generating data on both human face datasets.
Fig. 10: Faces generated by U-Net generator trained on cropped CelebA. U-Net generator also generates artistic style images.
Moreover, we can see that the shapes of output images look more similar to the original ones, showing that the U-Net imposes
the constraints on shapes of the content and only change the texture of an image. To our surprising, the results look like statues,
which are made up of polyhedron.
Fig. 11: More generated samples on SCUT-FBP5500. Some of them look crazy but cool. The left faces are typical images from
FBP5500. The images on the right side are framed by round borders, which are probably from profile pictures. Although some
of the training images are not square, the Face2Anime CycleGAN still works well.
Fig. 12: Novel faces which are unseen during training. The generators are able to generate anime-style faces from unseen human
faces. This result shows the generalization of CycleGAN. However, some faces are only slightly changed in my experiment.
Fig. 13: Some samples of anime-to-human faces. Although it is quite difficult for a generator to generate real human faces from
anime faces. There are still some successful samples from the generated testing images. Figure 13 demonstrates some of them.
The generator from anime to real do learn some human face textures, such as smoother skin, smaller eyes, lower contrast hair
color, a straight nose, etc.
[1] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, [8] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A
Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Efros, “Unpaired image-to-image translation using
cycle-consistent adversarial networks,” arXiv preprint [21] Tim Salimans, Ian Goodfellow, Wojciech Zaremba,
arXiv:1703.10593, 2017. Vicki Cheung, Alec Radford, and Xi Chen, “Improved
techniques for training gans,” in Advances in Neural
[9] Taeksoo Kim, Moonsu Cha, Hyunsoo Kim, Jungkwon Information Processing Systems, 2016, pp. 2234–2242.
Lee, and Jiwon Kim, “Learning to discover cross-
domain relations with generative adversarial networks,” [22] Geoffrey E Hinton, Nitish Srivastava, Alex Krizhevsky,
arXiv preprint arXiv:1703.05192, 2017. Ilya Sutskever, and Ruslan R Salakhutdinov, “Improv-
ing neural networks by preventing co-adaptation of fea-
[10] Zili Yi, Hao Zhang, Ping Tan, and Minglun Gong, “Du- ture detectors,” arXiv preprint arXiv:1207.0580, 2012.
algan: Unsupervised dual learning for image-to-image
translation,” arXiv preprint, 2017. [23] Sergey Ioffe and Christian Szegedy, “Batch nor-
malization: Accelerating deep network training by
[11] Yaniv Taigman, Adam Polyak, and Lior Wolf, “Un- reducing internal covariate shift,” arXiv preprint
supervised cross-domain image generation,” arXiv arXiv:1502.03167, 2015.
preprint arXiv:1611.02200, 2016.
[24] Dmitry Ulyanov, Andrea Vedaldi, and Victor S. Lempit-
[12] Yanghua Jin, Jiakai Zhang, Minjun Li, Yingtao Tian, sky, “Instance normalization: The missing ingredient
Huachun Zhu, and Zhihao Fang, “Towards the au- for fast stylization,” CoRR, vol. abs/1607.08022, 2016.
tomatic anime characters creation with generative ad-
versarial networks,” arXiv preprint arXiv:1708.05509,
2017.
[17] Ziwei Liu, Ping Luo, Xiaogang Wang, and Xiaoou Tang,
“Deep learning face attributes in the wild,” in Proceed-
ings of International Conference on Computer Vision
(ICCV), 2015.