Cooperation between ELLIS Unit Madrid members

On Wednesday November 22nd, Daniel Hernández-Lobato Professor at Universidad Autónoma de Madrid (UAM) presented his work “Variational linearized Laplace approximation for Bayesian deep learning” in the research seminar series of the Signal Processing Group GTS – Universidad Carlos III de Madrid (UC3M)

More Info:

https://www.researchgate.net/publication/368829898_Variational_Linearized_Laplace_Approximation_for_Bayesian_Deep_Learning

Abstract:
Pre-trained deep neural networks can be adapted to perform uncertainty estimation by transforming them into Bayesian neural networks via methods such as Laplace approximation (LA) or its linearized form (LLA), among others. To make these methods more tractable, the generalized Gauss-Newton (GGN) approximation is often used. However, due to complex inefficiency difficulties, both LA and LLA rely on further approximations, such as Kronecker-factored or diagonal approximate GGN matrices, which can affect the results. To address these issues, we propose a new method for scaling LLA using a variational sparse Gaussian Process (GP) approximation based on the dual RKHS of GPs. Our method retains the predictive mean of the original model while allowing for efficient stochastic optimization and scalability in both the number of parameters and the size of the training dataset. Moreover, its training cost is independent of the number of training points, improving over previously existing methods. Our preliminary experiments indicate that it outperforms already existing efficient variants of LLA, such as accelerated LLA (ELLA), based on the Nystr\”om approximation.