Skip to content

Sampling with gradient-based Markov Chain Monte Carlo approaches

License

Notifications You must be signed in to change notification settings

alisiahkoohi/Langevin-dynamics

Repository files navigation

Sampling with gradient-based Markov Chain Monte Carlo approaches

PyTorch implementation of stochastic gradient Langevin dynamics (SGLD) and preconditioned SGLD (pSGLD), involving simple examples of using unadjusted Langevin dynamics and Metropolis-adjusted Langevin algorithm (MALA) to sample from a 2D Gaussian distribution and "banana" distribution.

Prerequisites

Follow the steps below to install the necessary libraries:

pip install langevin-sampling

Script descriptions

langevin_sampling/SGLD.py: SGLD sampler.

langevin_sampling/precondSGLD.py: pSGLD sampler.

langevin_sampling/samplers.py:

  • Implements LangevinDynamics class that given negative-log of unnormalized density function and starting guess, runs Langevin dynamics to sample from the given density.

  • Implements MetropolisAdjustedLangevin class that given negative-log of unnormalized density function and starting guess, runs MALA to sample from the given density.

examples/gaussian_LD.py: Sampling from a toy 2D Gaussian distribution with unadjusted Langevin dynamics.

examples/gaussian_MALA.py: Sampling from a toy 2D Gaussian distribution with MALA.

examples/banana_LD.py: Sampling from a toy "banana" distribution with unadjusted Langevin dynamics.

examples/banana_MALA.py: Sampling from a toy "banana" distribution with MALA.

Running the code

To sample a toy 2D Gaussian distribution with unadjusted Langevin dynamics, run:

python examples/gaussian_LD.py

To sample a toy 2D Gaussian distribution with MALA, run:

python examples/gaussian_MALA.py

To sample a toy "banana" distribution with unadjusted Langevin dynamics, run:

python examples/banana_LD.py

To sample a toy "banana" distribution with MALA, run:

python examples/banana_MALA.py

References

  • Max Welling and Yee Whye Teh. Bayesian Learning via Stochastic Gradient Langevin Dynamics. In Proceedings of the 28th International Conference on Machine Learning, ICML’11, pages 681–688, Madison, WI, USA, 2011. Omnipress. ISBN 9781450306195. doi: 10.5555/3104482.3104568. URL https://dl.acm.org/doi/abs/10.5555/3104482.3104568.

  • Chunyuan Li, Changyou Chen, David Carlson, and Lawrence Carin. Preconditioned Stochastic Gradient Langevin Dynamics for Deep Neural Networks. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, AAAI’16, pages 1788–1794. AAAI Press, 2016. doi: 10.5555/3016100.3016149. URL https://dl.acm.org/doi/abs/10.5555/3016100.3016149.

  • Roberts, G.O., Stramer, O. Langevin Diffusions and Metropolis-Hastings Algorithms. Methodology and Computing in Applied Probability 4, pages 337–357, 2002. doi: 10.1023/A:1023562417138. URL https://link.springer.com/article/10.1023/A:1023562417138.

  • Pagani F, Wiegand M, Nadarajah S. An n-dimensional Rosenbrock distribution for MCMC testing. arXiv preprint arXiv:1903.09556. 2019 Mar 22. URL https://arxiv.org/abs/1903.09556.

Questions

Please contact alisk@rice.edu for further questions.

Author

Ali Siahkoohi

About

Sampling with gradient-based Markov Chain Monte Carlo approaches

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages