Abstract
A crucial requirement for reliable deployment of deep learning models for safety-critical applications is the ability to identify out-of-distribution (OOD) data points, samples which differ from the training data and on which a model might underper-form. Previous work has attempted to tackle this problem using uncertainty estimation techniques. However, there is empirical evidence that a large family of these techniques do not detect OOD reliably in classification tasks.
This paper gives a theoretical explanation for said experimental findings and illustrates it on synthetic data. We prove that such techniques are not able to reliably identify OOD samples in a classification setting, since their level of confidence is generalized to unseen areas of the feature space. This result stems from the interplay between the representation of ReLU networks as piece-wise affine transformations, the saturating nature of activation functions like softmax, and the most widely-used uncertainty metrics.
This paper gives a theoretical explanation for said experimental findings and illustrates it on synthetic data. We prove that such techniques are not able to reliably identify OOD samples in a classification setting, since their level of confidence is generalized to unseen areas of the feature space. This result stems from the interplay between the representation of ReLU networks as piece-wise affine transformations, the saturating nature of activation functions like softmax, and the most widely-used uncertainty metrics.
| Original language | English |
|---|---|
| Conference proceedings | Proceedings of Machine Learning Research |
| Volume | 161 |
| Pages (from-to) | 1766-1776 |
| ISSN | 2640-3498 |
| Publication status | Published - 2021 |
| Event | Conference on Uncertainty in Artificial Intelligence - VIRTUAL Duration: 27 Jul 2021 → 30 Jul 2021 Conference number: 37 https://www.auai.org/uai2021/ |
Conference
| Conference | Conference on Uncertainty in Artificial Intelligence |
|---|---|
| Number | 37 |
| City | VIRTUAL |
| Period | 27/07/2021 → 30/07/2021 |
| Internet address |
Keywords
- Out-of-Distribution Detection
- Deep Learning
- ReLU Networks
- Uncertainty Estimation
- Safety-critical Applications
Fingerprint
Dive into the research topics of 'Know Your Limits: Uncertainty Estimation with ReLU Classifiers Fails at Reliable OOD Detection'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver