Mind the Gap when Conditioning Amortised Inference in Sequential Latent-Variable Models

Justin Bayer · Maximilian Soelch · Atanas Mirchev · Baris Kayalibay · Patrick van der Smagt

Keywords: [ amortized inference ] [ state-space models ] [ recurrent networks ] [ variational inference ]

[ Abstract ]
[ Paper ]
Thu 6 May 1 a.m. PDT — 3 a.m. PDT


Amortised inference enables scalable learning of sequential latent-variable models (LVMs) with the evidence lower bound (ELBO). In this setting, variational posteriors are often only partially conditioned. While the true posteriors depend, e.g., on the entire sequence of observations, approximate posteriors are only informed by past observations. This mimics the Bayesian filter---a mixture of smoothing posteriors. Yet, we show that the ELBO objective forces partially-conditioned amortised posteriors to approximate products of smoothing posteriors instead. Consequently, the learned generative model is compromised. We demonstrate these theoretical findings in three scenarios: traffic flow, handwritten digits, and aerial vehicle dynamics. Using fully-conditioned approximate posteriors, performance improves in terms of generative modelling and multi-step prediction.

Chat is not available.