Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Andrei Bursuc
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Andrei Bursuc
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Deep Learning is great:
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Deep Learning is great:
... but has several problems
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Covering this diversity with (sufficient) data and labels is highly challenging
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Why should I care about uncertainty?
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
In May 2016, there was the first fatality from an assisted driving system, caused by the perception system confusing the white side of a trailer for bright sky.
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
An image classification system erroneously identifies two African Americans as gorillas, raising concerns of racial discrimination.
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
What do we mean by uncertainty?
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Return a distribution over predictions instead of a single prediction:
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Good uncertainty estimates tell us when we can trust the predictions of our model.
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
I.I.D ptest(x,y)=ptrain(x,y)
(I.I.D. = Indepedent and Identically Distributed)
O.O.D ptest(x,y)≠ ptrain(x,y)
Examples of dataset shift:
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
I.I.D ptest(x,y)=ptrain(x,y)
(I.I.D. = Indepedent and Identically Distributed)
O.O.D ptest(x,y)≠ ptrain(x,y)
Examples of dataset shift:
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Samples from ImageNet-C
D. Hendrycks & T. Dietterich, Benchmarking Neural Network Robustness to Common Corruptions and Perturbations, ICLR 2019
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Corruption types for ImageNet-C
D. Hendrycks & T. Dietterich, Benchmarking Neural Network Robustness to Common Corruptions and Perturbations, ICLR 2019
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Y. Ovadia et al., Can You Trust Your Model's Uncertainty? Evaluating Predictive Uncertainty Under Dataset Shift, NeurIPS 2019
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Example images where model assigns >99.5% confidence
A. Nguyen et al., Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images, CVPR 2015
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
J.Z. Liu et al., Simple and Principled Uncertainty Estimation with Deterministic Deep Learning via Distance Awareness, arXiv 2020
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Calibration Error=∣predicted probability of correctnessConfidence−observed frequency of correctnessAccuracy∣
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Calibration of weather forecasts
Nate Silver, The singal and the noise
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Most neural networks output probability distributions, e.g., over object categories. Are these calibrated?
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
ECE=b=1∑bNnb∣acc(b)−conf(b)∣
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
C. Guo et al., On Calibration of Modern Neural Networks, ICML 2017
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
The effect of network depth (far left), width (middle left), Batch Normalization (middle right), and weight decay (far right) on miscalibration, as measured by ECE (lower is better).
C. Guo et al., On Calibration of Modern Neural Networks, ICML 2017
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
The effect of network depth (far left), width (middle left), Batch Normalization (middle right), and weight decay (far right) on miscalibration, as measured by ECE (lower is better).
C. Guo et al., On Calibration of Modern Neural Networks, ICML 2017
We kind of got too good at training these beasts
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Healthcare: model uncertainty for trusting the model or calling doctor; reject low-quality inputs
Chatbots: detect unknown sentences
Active Learning: use model uncertainty to decide which training examples are worth labeling
Bayesian Optimization: optimize an expensive black-box function by finding which configurations to explore next
Reinforcement Learning: use uncertainty for exploration vs. exploitation trade-off
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Autonomous vehicles: dataset shift: location, weather, time of day; use model uncertainty to decide when to trust model or hand-over to human
Healthcare: model uncertainty for trusting the model or calling doctor; reject low-quality inputs
Chatbots: detect unknown sentences
Active Learning: use model uncertainty to decide which training examples are worth labeling
Bayesian Optimization: optimize an expensive black-box function by finding which configurations to explore next
Reinforcement Learning: use uncertainty for exploration vs. exploitation trade-off
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Autonomous vehicles: dataset shift: location, weather, time of day; use model uncertainty to decide when to trust model or hand-over to human
Healthcare: model uncertainty for trusting the model or calling doctor; reject low-quality inputs
Chatbots: detect unknown sentences
Active Learning: use model uncertainty to decide which training examples are worth labeling
Bayesian Optimization: optimize an expensive black-box function by finding which configurations to explore next
Reinforcement Learning: use uncertainty for exploration vs. exploitation trade-off
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Autonomous vehicles: dataset shift: location, weather, time of day; use model uncertainty to decide when to trust model or hand-over to human
Healthcare: model uncertainty for trusting the model or calling doctor; reject low-quality inputs
Chatbots: detect unknown sentences
Active Learning: use model uncertainty to decide which training examples are worth labeling
Bayesian Optimization: optimize an expensive black-box function by finding which configurations to explore next
Reinforcement Learning: use uncertainty for exploration vs. exploitation trade-off
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Autonomous vehicles: dataset shift: location, weather, time of day; use model uncertainty to decide when to trust model or hand-over to human
Healthcare: model uncertainty for trusting the model or calling doctor; reject low-quality inputs
Chatbots: detect unknown sentences
Active Learning: use model uncertainty to decide which training examples are worth labeling
Bayesian Optimization: optimize an expensive black-box function by finding which configurations to explore next
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Autonomous vehicles: dataset shift: location, weather, time of day; use model uncertainty to decide when to trust model or hand-over to human
Healthcare: model uncertainty for trusting the model or calling doctor; reject low-quality inputs
Chatbots: detect unknown sentences
Active Learning: use model uncertainty to decide which training examples are worth labeling
Bayesian Optimization: optimize an expensive black-box function by finding which configurations to explore next
Reinforcement Learning: use uncertainty for exploration vs. exploitation trade-off
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
There are two main types of uncertainties each with its own pecularities
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Problems caused by sensor quality, natural randomness, that cannot be explained by our data.
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Problems caused by sensor quality, natural randomness, that cannot be explained by our data.
Aleatoric / Data uncertainty
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Problems caused by sensor quality, natural randomness, that cannot be explained by our data.
Aleatoric / Data uncertainty
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Problems caused by sensor quality, natural randomness, that cannot be explained by our data.
Aleatoric / Data uncertainty
aleator (lat.) = dice player
cannot be reduced, but can be learned
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Similarly looking objects also fall into this category
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Similarly looking objects also fall into this category
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Credit: A. Malinin
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Credit: A. Malinin
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
In urban scenes this type of uncertainty is frequently caused by similarly-looking classes:
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Credit: A. Malinin
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Credit: A. Malinin
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Low entropy
High entropy
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
In layman words data uncertainty is called the: known unknown
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Lack of knowledge about the process that generated the data
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Lack of knowledge about the process that generated the data
Epistemic/Knowledge uncertainty
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Lack of knowledge about the process that generated the data
Epistemic/Knowledge uncertainty
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Lack of knowledge about the process that generated the data
Epistemic/Knowledge uncertainty
episteme (gr.) = knowledge
disappears given enough data
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Slide credit: Marcin Mozejko
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Image credit: Marcin Mozejko
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Let us consider a neural network model trained with several pictures of dog breeds.
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Let us consider a neural network model trained with several pictures of dog breeds.
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Let us consider a neural network model trained with several pictures of dog breeds.
Out-of-distribution uncertainty
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
In layman words, knowledge uncertainty is called the: unknown unknown
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Credit: A. Malinin
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Credit: A. Malinin
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Credit: A. Malinin
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
"Our model exhibits in (d) increased aleatoric uncertainty on object boundaries and for objects far from the camera. Epistemic uncertainty accounts for our ignorance about which model generated our collected data. In (e) our model exhibits increased epistemic uncertainty for semantically and visually challenging pixels. The bottom row shows a failure case of the segmentation model when the model fails to segment the footpath due to increased epistemic uncertainty, but not aleatoric uncertainty."
A. Kendall and Y. Gal, What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision?, NeurIPS 2017.
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Measuring the quality of the uncertainty can be challenging due to lack of ground truth, i.e., no “right answer” in some cases
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Dropout: A Simple Way to Prevent Neural Networks from Overfitting, Srivastava et al., JMLR 2014
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
>>> x = torch.full((3, 5), 1.0).requires_grad_() >>> xtensor([[ 1., 1., 1., 1., 1.], [ 1., 1., 1., 1., 1.], [ 1., 1., 1., 1., 1.]])>>> dropout = nn.Dropout(p = 0.75) >>> y = dropout(x)>>> ytensor([[ 0., 0., 4., 0., 4.], [ 0., 4., 4., 4., 0.], [ 0., 0., 4., 0., 0.]])>>> l = y.norm(2, 1).sum()>>> l.backward()>>> x.gradtensor([[ 0.0000, 0.0000, 2.8284, 0.0000, 2.8284] [ 0.0000, 2.3094, 2.3094, 2.3094, 0.0000] [ 0.0000, 0.0000, 4.0000, 0.0000, 0.0000]])
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
>>> x = torch.full((3, 5), 1.0).requires_grad_() >>> xtensor([[ 1., 1., 1., 1., 1.], [ 1., 1., 1., 1., 1.], [ 1., 1., 1., 1., 1.]])>>> dropout = nn.Dropout(p = 0.75) >>> y = dropout(x)>>> ytensor([[ 0., 0., 4., 0., 4.], [ 0., 4., 4., 4., 0.], [ 0., 0., 4., 0., 0.]])>>> l = y.norm(2, 1).sum()>>> l.backward()>>> x.gradtensor([[ 0.0000, 0.0000, 2.8284, 0.0000, 2.8284] [ 0.0000, 2.3094, 2.3094, 2.3094, 0.0000] [ 0.0000, 0.0000, 4.0000, 0.0000, 0.0000]])
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
For a given network
model = nn.Sequential(nn.Linear(10, 100), nn.ReLU(), nn.Linear(100, 50), nn.ReLU(), nn.Linear(50, 2));
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
For a given network
model = nn.Sequential(nn.Linear(10, 100), nn.ReLU(), nn.Linear(100, 50), nn.ReLU(), nn.Linear(50, 2));
model = nn.Sequential(nn.Linear(10, 100), nn.ReLU(), nn.Dropout(), nn.Linear(100, 50), nn.ReLU(), nn.Dropout(), nn.Linear(50, 2));
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
A model using dropout has to be set in train or test mode
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
A model using dropout has to be set in train or test mode
The method nn.Module.train(mode)
recursively sets the flag training
to
all sub-modules.
>>> dropout = nn.Dropout()>>> model = nn.Sequential(nn.Linear(3, 10), dropout, nn.Linear(10, 3)) >>> dropout.trainingTrue>>> model.train(False)Sequential ((0): Linear (3 -> 10) (1): Dropout (p = 0.5) (2): Linear (10 -> 3))>>> dropout.training False
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
A model using dropout has to be set in train or test mode
>>> dropout = nn.Dropout()>>> model = nn.Sequential(nn.Linear(3, 10), dropout, nn.Linear(10, 3)) >>> x = torch.full((1, 3), 1.0) >>> model.train()Sequential ((0): Linear (3 -> 10) (1): Dropout (p = 0.5) (2): Linear (10 -> 3))>>> model(x)tensor([[ 0.5360, -0.5225, -0.5129]], grad_fn=<ThAddmmBackward>)>>> model(x)tensor([[ 0.6134, -0.6130, -0.5161]], grad_fn=<ThAddmmBackward>)
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
A model using dropout has to be set in train or test mode
>>> dropout = nn.Dropout()>>> model = nn.Sequential(nn.Linear(3, 10), dropout, nn.Linear(10, 3)) >>> x = torch.full((1, 3), 1.0) >>> model.train()Sequential ((0): Linear (3 -> 10) (1): Dropout (p = 0.5) (2): Linear (10 -> 3))>>> model(x)tensor([[ 0.5360, -0.5225, -0.5129]], grad_fn=<ThAddmmBackward>)>>> model(x)tensor([[ 0.6134, -0.6130, -0.5161]], grad_fn=<ThAddmmBackward>)>>>>>> model.eval()Sequential ((0): Linear (3 -> 10) (1): Dropout (p = 0.5) (2): Linear (10 -> 3))>>> model(x)tensor([[ 0.5772, -0.0944, -0.1168]], grad_fn=<ThAddmmBackward>)>>> model(x)tensor([[ 0.5772, -0.0944, -0.1168]], grad_fn=<ThAddmmBackward>)
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Dropout as Bayesian approximation: representing model uncertainty in deep learning, Y. Gal, ICML 2016
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Image credit: Eric Ma
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Gal and Ghahramani build upon the ensembling view of Dropout and show that when training a network with dropout with a standard classification or regression objective, one is actually implicitly doing variational inference to match the posterior distribution of the weights.
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Proper epistemic uncertainty estimates at x can be obtained in a principled way using Monte-Carlo integration:
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Yarin Gal's demo.
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
class SimpleModel(nn.Module): def __init__(self, p, decay): super(SimpleModel, self).__init__() self.dropout_p = p self.decay = decay self.f = nn.Sequential( nn.Linear(1,20), nn.ReLU(), nn.Dropout(p=self.dropout_p), nn.Linear(20, 20), nn.ReLU(), nn.Dropout(p=self.dropout_p), nn.Linear(20,1) ) def forward(self, X): return self.f(X)
def uncertainty_estimate(X, model, iters=200, l2=0.01): model.train() outputs = np.hstack([model(X[:, np.newaxis]).data.numpy() \ for i in range(iters)]) y_mean = outputs.mean(axis=1) y_variance = outputs.var(axis=1) tau = l2 * (1. - model.dropout_p) / (2. * N * model.decay) y_variance += (1. / tau) y_std = np.sqrt(y_variance) return y_mean, y_std
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Y. Gal, Dropout as Bayesian approximation: representing model uncertainty in deep learning, ICML 2016
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Y. Gal, Dropout as Bayesian approximation: representing model uncertainty in deep learning, ICML 2016
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
A. Kendall and Y. Gal, What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision?, NeurIPS 2017.
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Semantic Segmentation performance on CamVid
What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision?, A. Kendall and Y. Gal, NeurIPS 2017
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Monocular Depth Regression Performance
What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision?, A. Kendall and Y. Gal, NeurIPS 2017
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Aleatoric vs. Epistemic Uncertainty for Out of Dataset Examples
What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision?, A. Kendall and Y. Gal, NeurIPS 2017
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Multiple follow-up papers by Gal and friends:
Marc LELARGE and Andrei BURSUC | Deep Learning Do It Yourself | 15.2 Uncertainty estimation - MCDropout
Andrei Bursuc
Keyboard shortcuts
↑, ←, Pg Up, k | Go to previous slide |
↓, →, Pg Dn, Space, j | Go to next slide |
Home | Go to first slide |
End | Go to last slide |
Number + Return | Go to specific slide |
b / m / f | Toggle blackout / mirrored / fullscreen mode |
c | Clone slideshow |
p | Toggle presenter mode |
t | Restart the presentation timer |
?, h | Toggle this help |
Esc | Back to slideshow |