date | tags |
---|---|
2020-06-15 |
paper, deep-learning, uncertainty |
Balaji Lakshminarayanan, Alexander Pritzel, Charles Blundell
Neural Information Processing Systems (NIPS 2017)
Year: 2017
Github: https://paperswithcode.com/paper/simple-and-scalable-predictive-uncertainty
The current work explores the idea of using deep learning bagging-based ensembles to estimate the uncertainty of a prediction. The main idea consists of training
The authors suggest the possibility of using adversarial training in addition to the ensembling mechanism to enhance the diversity of the data. Adversarial training means adding to a given input sample a small portion of the sign of the gradient of the loss wrt that sample:
The authors choose to use Negative Log Likelihood for regression problems because they assume that the output of the model is Gaussian, hence they can calculate probabilities using a PDF. A benefit of this metric over MSE or RMSE is that it takes into account the uncertainty of the estimation.
The results show potential benefits of the current methodology vs the MC Dropout method, as shown in the following figures.