What Uncertainties Do We Need to capture in Deep Learning? [with code]
Ibrahim Sobh - PhD
?? Senior Expert of Artificial Intelligence, Valeo Group | LinkedIn Top Voice | Machine Learning | Deep Learning | Data Science | Computer Vision | NLP | Developer | Researcher | Lecturer
There are two major types of uncertainty one can model
Understanding what a model does not know is a critical part of many machine learning systems
Bayesian deep learning can capture:
Aleatoric and epistemic uncertainty models are not mutually exclusive. The combination is able to achieve new state-of-the-art results.
Aleatoric uncertainty can further be categorized into:?
For example, Homoscedastic regression assumes constant observation noise σ for every input point x. Heteroscedastic regression, on the other hand, assumes that observation noise can vary with input x.
Approaches:
To capture epistemic uncertainty in a neural network (NN) we put a prior distribution over its weights, for example a Gaussian prior distribution: W samples from N (0, I). Bayesian neural networks replace the deterministic network’s weight parameters with distributions over these parameters, and instead of optimizing the network weights directly we average over all possible weights (marginalization). Bayesian inference is used to compute the posterior over the weights p(W|X, Y). This posterior captures the set of plausible model parameters, given the data. We can captured model uncertainty by approximating the distribution p(W|X, Y) via Dropout variational inference, a practical approach for approximate inference in large and complex models. This inference is done by training a model with dropout before every weight layer, and by also performing dropout at test time to sample from the approximate posterior (Monte Carlo dropout).
We can model Heteroscedastic aleatoric uncertainty just by changing our loss functions. Because this uncertainty is a function of the input data, we can learn to predict it using a deterministic mapping from inputs to model outputs. For example in regression, the model predicts not only a mean vale?y^?but also a variance?σ2. Similarly, Homoscedastic aleatoric uncertainty can be modeled in a similar way, but the uncertainty parameter will no longer be a model output, but a free parameter we optimize.
Conclusion?
领英推荐
??Summary
?? Epistemic uncertainty:?
???Aleatoric uncertainty:
References:
Thank you
Professor at Firat University
5 年A Brief Survey and an Application of Semantic Image Segmentation for Autonomous Driving
Professor at Firat University
5 年https://arxiv.org/pdf/1808.08413.pdf