STOCHASTIC GRADIENT LANGEVIN DYNAMICS THAT EXPLOIT NEURAL NETWORK STRUCTURE
Zachary Nado ⋅ Jasper Snoek ⋅ Roger Grosse ⋅ David Duvenaud ⋅ James Martens ⋅ Bowen Xu
Abstract
Tractable approximate Bayesian inference for deep neural networks remains challenging. Stochastic Gradient Langevin Dynamics (SGLD) offers a tractable approximation to the gold standard of Hamiltonian Monte Carlo. We improve on existing methods for SGLD by incorporating a recently-developed tractable approximation of the Fisher information, known as K-FAC, as a preconditioner.
Video
Chat is not available.
Successful Page Load