Skip to content

Implementation of a Variational Auto-Encoder in TensorFlow

License

Notifications You must be signed in to change notification settings

TamerAbdElaziz/VAE-TensorFlow

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

##Variational Auto-encoder

This is an improved implementation of the paper Stochastic Gradient VB and the Variational Auto-Encoder by D. Kingma and Prof. Dr. M. Welling. This code uses ReLUs and the adam optimizer, instead of sigmoids and adagrad. These changes make the network converge much faster.

I also created a Theano and a Torch version.

To run the MNIST experiment:

python main.py

###NB: This code is not as nicely polished as the Torch7 and Theano version. It is mainly for playing around with TensorFlow, which is why I tried to add as many of its bells and whistles as possible. PRs to make it more "TensorFlowy" are welcomed! Specifically if I made a mistake that causes a slow down.

There is no continuous version for now, but there will probably be one in the near future.

The code is MIT licensed.

About

Implementation of a Variational Auto-Encoder in TensorFlow

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%