Skip to content

Latest commit

 

History

History
17 lines (9 loc) · 948 Bytes

README.MD

File metadata and controls

17 lines (9 loc) · 948 Bytes

##Variational Auto-encoder

This is an improved implementation of the paper Stochastic Gradient VB and the Variational Auto-Encoder by D. Kingma and Prof. Dr. M. Welling. This code uses ReLUs and the adam optimizer, instead of sigmoids and adagrad. These changes make the network converge much faster.

I also created a Theano and a Torch version.

To run the MNIST experiment:

python main.py

###NB: This code is not as nicely polished as the Torch7 and Theano version. It is mainly for playing around with TensorFlow, which is why I tried to add as many of its bells and whistles as possible. PRs to make it more "TensorFlowy" are welcomed! Specifically if I made a mistake that causes a slow down.

There is no continuous version for now, but there will probably be one in the near future.

The code is MIT licensed.