+ All Categories
Home > Documents > Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2...

Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2...

Date post: 03-Jan-2021
Category:
Upload: others
View: 3 times
Download: 0 times
Share this document with a friend
22
1 MIT 2 IIIS, Tsinghua University 3 Adobe Research 4 CMU Differentiable Augmentation for Data-Efficient GAN Training NeurIPS 2020 Shengyu Zhao 1,2 Zhijian Liu 1 Ji Lin 1 Song Han 1 Jun-Yan Zhu 3,4
Transcript
Page 1: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

1MIT 2IIIS, Tsinghua University 3Adobe Research 4CMU

Differentiable Augmentationfor Data-Efficient GAN Training

NeurIPS 2020

Shengyu Zhao1,2 Zhijian Liu1 Ji Lin1 Song Han1Jun-Yan Zhu3,4

Page 2: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Computation AlgorithmComputation Algorithm

Data Is Expensive

FFHQ dataset: 70,000 selective post-processed human faces

Months or even years to collect the data,

along with prohibitive annotation costs.

ImageNet dataset: millions of images from diverse categories

Big Data

Page 3: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Sometimes Not Even Possible

3

Page 4: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Sometimes Not Even Possible

4

“The aim of the new directorate is to support fundamental scientific research ― with specific goals in mind. This is

not about solving incremental technical problems. As one example, in artificial intelligence, the focus would not be

on further refining current algorithms, but rather on developing profoundly new approaches that would enable

machines to "learn" using much smaller data sets ― a fundamental advance that would eliminate the need to

access immense data sets. Success in this work would have a double benefit: seeding economic benefits for the

U.S. while reducing the pressure to weaken privacy and civil liberties in pursuit of more "training" data.”

― L. Rafael Reif

Page 5: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

GANs Heavily Deteriorate Given Limited Data

5

Generated samples of StyleGAN2. The quality is poor given limited data.

Page 6: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Ob

am

a

100

imag

es

Generated samples of StyleGAN2 (Karras et al.)

using only hundreds of images

Ca

t (Sim

ard

et a

l.)

160

imag

es

Do

g (S

imard

et a

l.)

389

imag

es

GANs Heavily Deteriorate Given Limited Data

Page 7: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

GANs Heavily Deteriorate Given Limited Data

11.1

23.1

36.0

0

5

10

15

20

25

30

35

40

100% training data 20% training data 10% training data

FID↓

StyleGAN2 (baseline) + DiffAugment (ours)

CIFAR-10

Page 8: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Discriminator Overfitting

Page 9: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Data Augmentation

9

Data augmentation: enlarge datasets without collecting new samples.

Page 10: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

10

How to Augment GANs?

Page 11: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

#1 Approach: Augment reals only

Augment reals only: the same artifacts appear on the generated images.

Artifacts from Color jittering

Artifacts from Translation

Artifacts from Cutout (DeVries et al.)

Generated images

Page 12: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Augment 𝑫 only: the unbalanced optimization cripples training.

#2 Approach: Augment reals & fakes for 𝑫 only

Page 13: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Our approach (DiffAugment): Augment reals + fakes for both 𝐷 and 𝐺

#3 Approach: Differentiable Augmentation (Ours)

Color

Translation

Cutout

Color

Translation

Cutout

fakes reals

Page 14: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

11.1

23.1

36.0

9.9 12.2

14.5

0

5

10

15

20

25

30

35

40

100% training data 20% training data 10% training data

FID↓

StyleGAN2 (baseline) + DiffAugment (ours)

11.1

23.1

36.0

9.9 12.2

14.5

0

5

10

15

20

25

30

35

40

100% training data 20% training data 10% training data

FID↓

StyleGAN2 (baseline) + DiffAugment (ours)

Our Results

CIFAR-10

Page 15: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

ImageNet Generation (25% training data)

15

Page 16: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Low-Shot Generation

Ob

am

a

100

imag

es

Ca

t (Sim

ard

et a

l.)

160

imag

es

Do

g (S

imard

et a

l.)

389

imag

es

Page 17: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

100-Shot Generation

Generated samples of StyleGAN2 (baseline)

Generated samples of StyleGAN2 + DiffAugment (ours)

Page 18: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

0

10

20

30

40

50

60

Performance

FID

Scale/Shift (Noguchi et al.) MineGAN (Wang et al.) TransferGAN (Wang et al.) FreezeD (Mo et al.) Ours

1

10

100

1000

10000

100000

Data

# T

rain

ing Im

ages

Fine-Tuning vs. Ours

No pre-training

100-shot Obama

Page 19: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Fine-Tuning vs. Ours

19

TransferGAN (a state-of-the-art fine-tuning method)

70,000 FFHQ faces + 100 Obama portraits

Ours

only 100 Obama portraits

Page 20: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

100-Shot Interpolation

The smooth interpolation results suggest little overfitting of our method even given only 100 imagesof Obama, grumpy cat, panda, the Bridge of Sighs, the Medici Fountain, the Temple of Heaven, and Wuzhen.

Page 21: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Data-Efficient Deep Learning

21

Various factual and ethical reasons could cause limited data available.

This research will help alleviate these limitations.

Rare incidents Privacy concerns Under-represented subpopulations

Page 22: Differentiable Augmentation for Data-Efficient GAN Training · 100 s Generated samples of StyleGAN2 (Karras et al.) using only hundreds of images Cat l.) 160 s Dog l.) 389 s GANs

Thanks for listening!

22

Our code, datasets, and models are publicly available at

https://github.com/mit-han-lab/data-efficient-gans.


Recommended