Reading Note: Progressive Growing of GANs for Improved Quality, Stability, and Variation

42 篇文章 0 订阅
2 篇文章 0 订阅

TITLE: Progressive Growing of GANs for Improved Quality, Stability, and Variation

AUTHOR: Tero Karras, Timo Aila, Samuli Laine, Jaakko Lehtinen

ASSOCIATION: NVIDIA

FROM: ICLR2018

CONTRIBUTION

A training methodology is proposed for GANs which starts with low-resolution images, and then progressively increases the resolution by adding layers to the networks. This incremental nature allows the training to first discover large-scale structure of the image distribution and then shift attention to increasingly finer scale detail, instead of having to learn
all scales simultaneously.

METHOD

PROGRESSIVE GROWING OF GANS

The following figure illustrates the training procedure of this work.

Framework

The training starts with both the generator G and discriminator D having a low spatial resolution of 4×4 pixels. As the training advances, successive layers are incrementally added to G and D, thus increasing the spatial resolution of the generated images. All existing layers remain trainable throughout the process. Here N×N refers to convolutional layers operating on N×N spatial resolution. This allows stable synthesis in high resolutions and also speeds up training considerably.

Framework

fade in is adopted when the new layers are added to double resolution of the generator G and discriminator D smoothly. This example illustrates the transition from 16×16 images (a) to 32×32 images (c). During the transition (b) the layers that operate on the higher resolution works like a residual block, whose weight α increases linearly from 0 to 1. Here 2x and 0.5x refer to doubling and halving the image resolution using nearest neighbor filtering and average pooling, respectively. The toRGB represents a layer that projects feature vectors to RGB colors and fromRGB does the reverse; both use 1×1 convolutions. When training the discriminator, the real images are downscaled to match the current resolution of the network. During a resolution transition, interpolation is carried out between two resolutions of the real images, similarly to how the generator output combines two resolutions.

INCREASING VARIATION USING MINIBATCH STANDARD DEVIATION

  1. Compute the standard deviation for each feature in each spatial location over the minibatch.
  2. Average these estimates over all features and spatial locations to arrive at a single value.
  3. Consturct one additional (constant) feature map by replicating the value and concatenate it to all spatial locations and over the minibatch

NORMALIZATION IN GENERATOR AND DISCRIMINATOR

EQUALIZED LEARNING RATE. A trivial N(0;1) initialization is used and then explicitly the weights are scaled at runtime. To be precise, w^i=wi/c , where wi are the weights and c is the per-layer normalization constant from He’s initializer.The benefit of doing this dynamically instead of during initialization is somewhat subtle, and relates to the scale-invariance in commonly used adaptive stochastic gradient descent methods.

PIXELWISE FEATURE VECTOR NORMALIZATION IN GENERATOR. To disallow the scenario where the magnitudes in the generator and discriminator spiral out of control as a result of competition, the feature vector is normalized in each pixel to unit length in the generator after each convolutional layer, using a variant of “local response normalization”, configured as

bx,y=ax,y/1Nj=0N1(ajx,y)2+ϵ

where ϵ=108 , N is the number of feature maps, and ax,y is original feature vector, bx,y is the normalized feature vector in pixel (x,y) .

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值