We describe a new training methodology for generative adversarial networks.
The key idea is to grow both the generator and discriminator progressively:
starting from a low resolution, we add new layers that model increasingly fine
details as training progresses. This both speeds the training up and greatly
stabilizes it, allowing us to produce images of unprecedented quality, e.g.,
CelebA images at 1024^2. We also propose a simple way to increase the variation
in generated images, and achieve a record inception score of 8.80 in
unsupervised CIFAR10. Additionally, we describe several implementation details
that are important for discouraging unhealthy competition between the generator
and discriminator. Finally, we suggest a new metric for evaluating GAN results,
both in terms of image quality and variation. As an additional contribution, we
construct a higher-quality version of the CelebA dataset.
Beschreibung
[1710.10196] Progressive Growing of GANs for Improved Quality, Stability, and Variation
%0 Journal Article
%1 karras2017progressive
%A Karras, Tero
%A Aila, Timo
%A Laine, Samuli
%A Lehtinen, Jaakko
%D 2017
%K generative-models readings stable
%T Progressive Growing of GANs for Improved Quality, Stability, and
Variation
%U http://arxiv.org/abs/1710.10196
%X We describe a new training methodology for generative adversarial networks.
The key idea is to grow both the generator and discriminator progressively:
starting from a low resolution, we add new layers that model increasingly fine
details as training progresses. This both speeds the training up and greatly
stabilizes it, allowing us to produce images of unprecedented quality, e.g.,
CelebA images at 1024^2. We also propose a simple way to increase the variation
in generated images, and achieve a record inception score of 8.80 in
unsupervised CIFAR10. Additionally, we describe several implementation details
that are important for discouraging unhealthy competition between the generator
and discriminator. Finally, we suggest a new metric for evaluating GAN results,
both in terms of image quality and variation. As an additional contribution, we
construct a higher-quality version of the CelebA dataset.
@article{karras2017progressive,
abstract = {We describe a new training methodology for generative adversarial networks.
The key idea is to grow both the generator and discriminator progressively:
starting from a low resolution, we add new layers that model increasingly fine
details as training progresses. This both speeds the training up and greatly
stabilizes it, allowing us to produce images of unprecedented quality, e.g.,
CelebA images at 1024^2. We also propose a simple way to increase the variation
in generated images, and achieve a record inception score of 8.80 in
unsupervised CIFAR10. Additionally, we describe several implementation details
that are important for discouraging unhealthy competition between the generator
and discriminator. Finally, we suggest a new metric for evaluating GAN results,
both in terms of image quality and variation. As an additional contribution, we
construct a higher-quality version of the CelebA dataset.},
added-at = {2019-10-18T14:26:35.000+0200},
author = {Karras, Tero and Aila, Timo and Laine, Samuli and Lehtinen, Jaakko},
biburl = {https://www.bibsonomy.org/bibtex/293428a00c0815fcfe604bb733c0ff808/kirk86},
description = {[1710.10196] Progressive Growing of GANs for Improved Quality, Stability, and Variation},
interhash = {4b5a080433b88e6f93ea4cfd9409cc60},
intrahash = {93428a00c0815fcfe604bb733c0ff808},
keywords = {generative-models readings stable},
note = {cite arxiv:1710.10196Comment: Final ICLR 2018 version},
timestamp = {2019-10-18T14:26:35.000+0200},
title = {Progressive Growing of GANs for Improved Quality, Stability, and
Variation},
url = {http://arxiv.org/abs/1710.10196},
year = 2017
}