Skip to content

Pytorch implementation of LARGE SCALE GAN TRAINING FOR HIGH FIDELITY NATURAL IMAGE SYNTHESIS (BigGAN)

License

Notifications You must be signed in to change notification settings

sxhxliang/BigGAN-pytorch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

be65631 · Feb 23, 2019

History

12 Commits
Nov 15, 2018
Dec 4, 2018
Nov 16, 2018
Nov 14, 2018
Nov 13, 2018
Feb 23, 2019
Nov 15, 2018
Nov 15, 2018
Nov 14, 2018
Nov 15, 2018
Nov 15, 2018
Nov 15, 2018
Nov 13, 2018
Nov 15, 2018
Nov 15, 2018

Repository files navigation

BigGAN-PyTorch

Pytorch implementation of LARGE SCALE GAN TRAINING FOR HIGH FIDELITY NATURAL IMAGE SYNTHESIS (BigGAN)

train imagenet

for 128*128*3 resolution

python main.py --batch_size 64  --dataset imagenet --adv_loss hinge --version biggan_imagenet --image_path /data/datasets

python main.py --batch_size 64  --dataset lsun --adv_loss hinge --version biggan_lsun --image_path /data1/datasets/lsun/lsun

python main.py --batch_size 64  --dataset lsun --adv_loss hinge --version biggan_lsun --parallel True --gpus 0,1,2,3 --use_tensorboard True

Different

  • not use cross-replica BatchNorm (Ioffe & Szegedy, 2015) in G

Compatability

  • CPU
  • GPU

Pretrained Models

LSUN Pretrained model Download

Some methods in the paper to avoid model collapse, please see the paper and retrain your model.

Performance

  • Infact, as mentioned in the paper, the model will collapse
  • I use LSUN datasets to train this model maybe cause bad performance due to the class of classroom is more complex than �ImageNet

Results

LSUN DATASETS(two classes): classroom and church_outdoor

  • iter 82200 (128x128) batch_size 64
  • iter 128200
  • iter 365000
  • iter 800000
  • iter 900000

About

Pytorch implementation of LARGE SCALE GAN TRAINING FOR HIGH FIDELITY NATURAL IMAGE SYNTHESIS (BigGAN)

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages