Variational AutoEncoder


Table of Contents

Introduction

This is an Pytorch implementation of Variational AutoEncoder, model structure was inspired by the one used with Stable-Diffusion.

Performance

The following is the result of training on the ImageNet-1K dataset with 256x256 resolution and 64 latent size for 55100 steps with batch size of 32. The final L1 loss is 0.015656. training_example

Checkpoints

Checkpoints are in the sub-directory checkpoints, and are in the format of DATASET_IMAGESIZE_LATENTSIZE_STEPS-BATCHSIZE_LOSS.pt

License

vae is distributed under the terms of the MIT license.