pith. machine review for the scientific record. sign in

arxiv: 1505.05424 · v2 · submitted 2015-05-20 · 📊 stat.ML · cs.LG

Recognition: unknown

Weight Uncertainty in Neural Networks

Charles Blundell, Daan Wierstra, Julien Cornebise, Koray Kavukcuoglu

Authors on Pith no claims yet
classification 📊 stat.ML cs.LG
keywords uncertaintyweightslearningneuralprincipledusedweightalgorithm
0
0 comments X
read the original abstract

We introduce a new, efficient, principled and backpropagation-compatible algorithm for learning a probability distribution on the weights of a neural network, called Bayes by Backprop. It regularises the weights by minimising a compression cost, known as the variational free energy or the expected lower bound on the marginal likelihood. We show that this principled kind of regularisation yields comparable performance to dropout on MNIST classification. We then demonstrate how the learnt uncertainty in the weights can be used to improve generalisation in non-linear regression problems, and how this weight uncertainty can be used to drive the exploration-exploitation trade-off in reinforcement learning.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 2 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Concrete Problems in AI Safety

    cs.AI 2016-06 accept novelty 7.0

    The paper categorizes five concrete AI safety problems arising from flawed objectives, costly evaluation, and learning dynamics.

  2. Portfolio Optimization Proxies under Label Scarcity and Regime Shifts via Bayesian and Deterministic Students under Semi-Supervised Sandwich Training

    cs.LG 2026-04 unverdicted novelty 4.0

    A semi-supervised teacher-student framework enables neural networks to proxy CVaR portfolio optimization using synthetic data augmentation for scarce labels and regime shifts.