Abstract
We investigate the problem of determining the predictive confidence (or, conversely, uncertainty) of a neural classifier through the lens of low-resource languages. By training models on sub-sampled datasets in three different languages, we assess the quality of estimates from a wide array of approaches and their dependence on the amount of available data. We find that while approaches based on pre-trained models and ensembles achieve the best results overall, the quality of uncertainty estimates can surprisingly suffer with more data. We also perform a qualitative analysis of uncertainties on sequences, discovering that a model's total uncertainty seems to be influenced to a large degree by its data uncertainty, not model uncertainty. All model implementations are open-sourced in a software package.
| Originalsprog | Engelsk |
|---|---|
| Titel | Findings of 2022 Conference on Empirical Methods in Natural Language Processing (EMNLP) |
| Publikationsdato | 7 dec. 2022 |
| DOI | |
| Status | Udgivet - 7 dec. 2022 |
| Begivenhed | Empirical Methods in Natural Language Processing - Abu Dhabi National Exhibition Center (ADNEC), Abu Dhabi, United Arab Emirates Varighed: 7 dec. 2022 → 11 dec. 2022 https://2022.emnlp.org/ |
Konference
| Konference | Empirical Methods in Natural Language Processing |
|---|---|
| Lokation | Abu Dhabi National Exhibition Center (ADNEC) |
| Land/Område | United Arab Emirates |
| By | Abu Dhabi |
| Periode | 07/12/2022 → 11/12/2022 |
| Internetadresse |
Emneord
- Predictive confidence
- Neural classifier
- Low-resource languages
- Uncertainty estimation
- Pre-trained models
- Ensembles
- Data uncertainty
- Model uncertainty
- Sequence analysis
- Open-source software