Implementing and optimizing two generative models from scratch for image generation (ICL):
- Variational Auto Encoder (VAE)
- Deep Convolutional Generatve Adversarial Network (DCGAN)
On MNIST dataset.
On CIFAR10 dataset
- DCGAN architecture adapted to the 32x32 size of the images
- Gradient penalty term for discriminator
- 5% of labels are flipped
- One-sided label smooting (Real labels set to 0.9 for stronger gradient signals)
- Batch normalization
- Data augmentation
- Learning rate scheduling
- Leaky ReLU in the discriminator (from literature: Radford et al. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks)