Bayesian neural networks (BNNs) have received an increased interest in the last years. In BNNs, a complete posterior distribution of the unknown weight and bias parameters of the network is produced during the training stage. This probabilistic estimation offers several advantages with respect to point-wise estimates, in particular, the ability to provide uncertainty quantification when predicting new data. This feature inherent to the Bayesian paradigm, is useful in countless machine learning applications. It is particularly appealing in areas where decision-making has a crucial impact, such as medical healthcare or autonomous driving. The main challenge of BNNs is the computational cost of the training procedure since Bayesian techniques often face a severe curse of dimensionality. Adaptive importance sampling (AIS) is one of the most prominent Monte Carlo methodologies benefiting from sounded convergence guarantees and ease for adaptation. This talk aims to show that AIS constitutes a successful approach for designing BNNs. More precisely, we introduce a novel algorithm PMCnet [1] that includes an efficient adaptation mechanism, exploiting geometric information on the complex (often multimodal) posterior distribution. Numerical results illustrate the excellent performance and the improved exploration capabilities of the proposed method for both shallow and deep neural networks.
(joint work with Y. Huang, V. Elvira, J.C. Pesquet)
[1] Y. Huang, E. Chouzenoux, V. Elvira and J.-C. Pesquet. Efficient Bayes Inference in Neural Networks through Adaptive Importance Sampling. To appear in Journal of Franklin Institute. 2023. https://arxiv.org/abs/2210.00993