Estimating the uncertainty of deep learning models has gained exponential interest in recent years, mainly due to the safety concerns for real applications. In this Arxiv, the latest literature from NIPS, ICML, ICLR, etc will be covered (on-updating). some papers may be included multiple times if they cover across topics.
- [Information Fusion 2021] A review of uncertainty quantification in deep learning: Techniques, applications, and challenges
- [arXiv2020] A Survey of Uncertainty in Deep Neural Networks
- [ACM Computing Surveys 2020] A Survey on Bayesian Deep Learning (github)
- [ACM Computing Surveys] A survey on concept drift adaptation (citation 2K+)
- [AIES2021] Uncertainty as a form of transparency: Measuring, communicating, and using uncertainty
- [arXiv2020] A Survey on Assessing the Generalization Envelope of Deep Neural Networks
- [arXiv2021] A Unified Survey on Anomaly, Novelty, Open-Set, and Out-of-Distribution Detection: Solutions and Future Challenges
- [NeurIPS2020] Tutorial: (Track2) Practical Uncertainty Estimation and Out-of-Distribution Robustness in Deep Learning
- [NeurIPS2021] Workshop: Bayesian Deep Learning
- [NeurPS2021] Workshop: Distribution shifts: connecting methods and applications (DistShift)
- I Can’t Believe It’s Not Better! (ICBINB) Workshop
- [A Survey on Evidential Deep Learning For Single-Pass Uncertainty Estimation]
- [npj Digital Medicine 2021] Second opinion needed: communicating uncertainty in medical machine learning
- [ICML2022] Workshop: Distribution-Free Uncertainty Quantification
- [Nature Machine Intelligence 2019] The need for uncertainty quantification in machine-assisted medical decision making
- [Nature Medicine 2023] Enhancing the reliability and accuracy of AI-enabled diagnosis via complementarity-driven deferral to clinicians
- [NeurIPS2019] Can you trust your model's uncertainty? Evaluating predictive uncertainty under dataset shift
- [ICLR2019] Benchmarking neural network robustness to common corruptions and perturbations
- [NeurIPS2021] Benchmarking Bayesian Deep Learning on Diabetic Retinopathy Detection Tasks
- [NeurIPS2021]Reliable and Trustworthy Machine Learning for Health Using Dataset Shift Detection
- [ICML2021]Evaluating Robustness of Predictive Uncertainty Estimation: Are Dirichlet-based Models Reliable?
- Openreview from ICLR
- [IJCAI2019]Statistical Guarantees for the Robustness of Bayesian Neural Networks (Robustness measurement)
- [ICML2022 Spotlight] On the Practicality of Deterministic Epistemic Uncertainty
- Performance of detecting correct/incorrect predictions
- uncertainty with increasing data shift
- [ICML2016] Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning (MCDropout)
- [NeurIPS2019] A simple baseline for bayesian uncertainty in deep learning (SWAG)
- [NeurIPS2019]Using Self-Supervised Learning Can Improve Model Robustness and Uncertainty
- [NeurIPS2020] Simple and Principled Uncertainty Estimation with Deterministic Deep Learning via Distance Awareness
- [NeurIPS2020] Liberty or Depth: Deep Bayesian Neural Nets Do Not Need Complex Weight Posterior Approximations
- [NeurIPS2020] Posterior Re-calibration for Imbalanced Datasets
- [ICLR2020] Your classifier is secretly an energy based model and you should treat it like one
- [ICASSP 2022] Uncertainty Estimation with a VAE-Classifier Hybrid Model
- [AAAI 2022] Transformer Uncertainty Estimation with Hierarchical Stochastic Attention
- [ICML2022] Transformer Neural Processes: Uncertainty-Aware Meta Learning Via Sequence Modeling
- [ICML2022] On the Practicality of Deterministic Epistemic Uncertainty
- [NeurIPS2022] Pitfalls of Epistemic Uncertainty Quantification through Loss Minimisation
- [NeurIPS2018] To Trust Or Not To Trust A Classifier (Trust Score)
- [ICLR2018] Training Confidence-calibrated Classifiers for Detecting Out-of-Distribution Samples (uniform distribution)
- [NeurIPS2019] Addressing Failure Prediction by Learning Model Confidence (ConfidNet)
- [NeurIPS2019] Accurate Layerwise Interpretable Competence Estimation (ALICE)
- [ICML2020] Confidence-Aware Learning for Deep Neural Networks (Correctness Ranking Loss)
- [NeurIPS2017] Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles -- proper loss function, Brier score, average aggrgation
- [NeurIPS2019] Accurate Uncertainty Estimation and Decomposition in Ensemble Learning
- [ICLR2021] Training independent subnetworks for robust prediction (MIMO)
- [NeurIPS2021] Scaling Ensemble Distribution Distillation to Many Classes with Proxy Targets (A Malinin and M Gales, Ensemble+Dirichlet)
- [ICML2017] On calibration of modern neural networks (T-Scaling)
- [NeurIPS2019] Verified Uncertainty Calibration
- [NeurIPS2020] Improving model calibration with accuracy versus uncertainty optimization
- [ICML2020] Mix-n-match: Ensemble and compositional methods for uncertainty calibration in deep learning
- [NeurIPS2020] Posterior Re-calibration for Imbalanced Datasets (a single hyper-parameter.)
- [NeurIPS2021] Revisiting the Calibration of Modern Neural Networks
- [NeurIPS2019] A simple baseline for bayesian uncertainty in deep learning (SWAG)
- [NeurIPS2021] On the Importance of Gradients for Detecting Distributional Shifts in the Wild (GradNorm)
- [NeurIPS2018] A Simple Unified Framework for Detecting Out-of-Distribution Samples and Adversarial Attacks (Mahalanobis Distance)
- [ICML2020]Uncertainty Estimation Using a Single Deep Deterministic Neural Network
- [UAI2020] Greedy Policy Search: A Simple Baseline for Learnable Test-Time Augmentation
- [MIDL2021] Test-Time Mixup Augmentation for Uncertainty Estimation in Skin Lesion Diagnosis
- [ICML2021]Evaluating Robustness of Predictive Uncertainty Estimation: Are Dirichlet-based Models Reliable?
- [NeurIPS2021] Reliable and Trustworthy Machine Learning for Health Using Dataset Shift Detection
- [Google 2021] Does Your Dermatology Classifier Know What It Doesn't Know? Detecting the Long-Tail of Unseen Conditions
- [Computer Methods and Programs in Biomedicine] Uncertainty quantification in DenseNet model using myocardial infarction ECG signals
- Dirichlet distribution
- ECG classification
- [npj Digital Medicine] Statistical uncertainty quantification to augment clinical decision support: a first implementation in sleep medicine
- Entropy of softmax
- EEG, human-in-the-loop
- [Neural Computing and Applications 2021]Leveraging the Bhattacharyya coefficient for uncertainty quantification in deep neural networks
- BC score
- CIFAR and HAM1000
- Class imbalance is noticed but not addressed
- [Nature Machine Intelligence 2019] The need for uncertainty quantification in machine-assisted medical decision making
- [Nature Medicine 2023] Enhancing the reliability and accuracy of AI-enabled diagnosis via complementarity-driven deferral to clinicians
- [NeurIPS2021]An Information-theoretic Approach to Distribution Shifts
- distribution shift = Covariate shift + Concept shift
- [NeurIPS2020] Why Normalizing Flows Fail to Detect Out-of-Distribution Data
- [NeurIPS2020] On the Value of Out-of-Distribution Testing: An Example of Goodhart's Law
- [NeurIPS2021] Exploring the Limits of Out-of-Distribution Detection
- [NeurIPS2021] A Winning Hand: Compressing Deep Networks Can Improve Out-of-Distribution Robustness
- [ICIR2017] A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks (Softmax)
- [ICLR2018] Enhancing The Reliability of Out-of-distribution Image Detection in Neural Networks (ODIN = T scaling + input perturbation)
- [NeurIPS2020] Certifiably Adversarially Robust Detection of Out-of-Distribution Data
- [NeurIPS202] Energy-based Out-of-distribution Detection (Energy score)
- [NeurIPS2021] On the Importance of Gradients for Detecting Distributional Shifts in the Wild (GradNorm)
- [NeurIPS2021] ReAct: Out-of-distribution Detection With Rectified Activations
- [NeurIPS2021] Single Layer Predictive Normalized Maximum Likelihood for Out-of-Distribution Detection
- [NeurIPS2021] Learning Causal Semantic Representation for Out-of-Distribution Prediction
- [NeurIPS2021] STEP: Out-of-Distribution Detection in the Presence of Limited In-Distribution Labeled Data
- [NeurIPS2021] Task-Agnostic Undesirable Feature Deactivation Using Out-of-Distribution Data
- [Normolising flows]
- [NeurIPS2018] Evidential deep learning to quantify classification uncertainty
- [NeurIPS2020] Posterior Network: Uncertainty Estimation without OOD Samples via Density-Based Pseudo-Counts [Codes]
- [ICML2021]Evaluating Robustness of Predictive Uncertainty Estimation: Are Dirichlet-based Models Reliable?
- [ICLR2022] Natural Posterior Network: Deep Bayesian Predictive Uncertainty for Exponential Family Distributions
- [AAAI2021] Multidimensional Uncertainty-Aware Evidential Neural Networks
- [ICASSP2022] SEED: SOUND EVENT EARLY DETECTION VIA EVIDENTIAL UNCERTAINTY
- [IEEE/CVF 2024] Evidential Uncertainty Quantification: A Variance-Based Perspective
- [NeurIPS2019] Likelihood Ratios for Out-of-Distribution Detection
- [ICLR2019] Do Deep Generative Models Know What They Don't Know?
- [NeurIPS2020] Likelihood Regret: An Out-of-Distribution Detection Score For Variational Auto-encoder
- [NeurIPS2021] Locally Most Powerful Bayesian Test for Out-of-Distribution Detection using Deep Generative Models
- [NeurIPS2018] Predictive Uncertainty Estimation via Prior Networks (A Malinin, M Gales)
- [video]
- [video]
- [NeurIPS2019] Reverse KL-Divergence Training of Prior Networks: Improved Uncertainty and Adversarial Robustness (A Malinin, M Gales)
- [NeurIPS2020] Towards Maximizing the Representation Gap between In-Domain & Out-of-Distribution Examples
- [NeurIPS2020] OOD-MAML: Meta-Learning for Few-Shot Out-of-Distribution Detection and Classification
- [NeurIPS2021] Characterizing Generalization under Out-Of-Distribution Shifts in Deep Metric Learning
- [NeurIPS2021] Towards a Theoretical Framework of Out-of-Distribution Generalization
- [NeurIPS2021] On the Out-of-distribution Generalization of Probabilistic Image Modelling (Cam&Huawei)
- [AAAI2023] Post-hoc Uncertainty Learning using a Dirichlet Meta-Model
- [[ICML2024] Decomposing Uncertainty for Large Language Models through Input Clarification Ensembling
- [ICLR2018] Enhancing The Reliability of Out-of-distribution Image Detection in Neural Networks (ODIN)
- [NeurIPS2018] A Simple Unified Framework for Detecting Out-of-Distribution Samples and Adversarial Attacks (Mahalanobis Distance)
- [ICML2020] Detecting out-of-distribution examples with gram matrices (Gram)
- [NeurIPS202] Energy-based Out-of-distribution Detection (Energy score)