Abstract
In Natural Language Generation (NLG) tasks,
for any input, multiple communicative goals are
plausible, and any goal can be put into words,
or produced, in multiple ways. We characterise
the extent to which human production varies
lexically, syntactically, and semantically across
four NLG tasks, connecting human production
variability to aleatoric or data uncertainty. We
then inspect the space of output strings shaped
by a generation system’s predicted probability
distribution and decoding algorithm to probe
its uncertainty. For each test input, we measure
the generator’s calibration to human production
variability. Following this instance-level ap-
proach, we analyse NLG models and decoding
strategies, demonstrating that probing a genera-
tor with multiple samples and, when possible,
multiple references, provides the level of detail
necessary to gain understanding of a model’s
representation of uncertainty.
for any input, multiple communicative goals are
plausible, and any goal can be put into words,
or produced, in multiple ways. We characterise
the extent to which human production varies
lexically, syntactically, and semantically across
four NLG tasks, connecting human production
variability to aleatoric or data uncertainty. We
then inspect the space of output strings shaped
by a generation system’s predicted probability
distribution and decoding algorithm to probe
its uncertainty. For each test input, we measure
the generator’s calibration to human production
variability. Following this instance-level ap-
proach, we analyse NLG models and decoding
strategies, demonstrating that probing a genera-
tor with multiple samples and, when possible,
multiple references, provides the level of detail
necessary to gain understanding of a model’s
representation of uncertainty.
Originalsprog | Engelsk |
---|---|
Titel | Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing |
Antal sider | 22 |
Udgivelsessted | Singapore |
Forlag | Association for Computational Linguistics |
Publikationsdato | dec. 2023 |
Sider | 14349–14371 |
DOI | |
Status | Udgivet - dec. 2023 |
Emneord
- Natural Language Generation
- Communicative goals
- Human production variability
- Aleatoric uncertainty
- Decoding algorithm
- Lexical variability
- Syntactic variability
- Semantic variability
- Model calibration
- Uncertainty representation
Fingeraftryk
Dyk ned i forskningsemnerne om 'What Comes Next? Evaluating Uncertainty in Neural Text Generators Against Human Production Variability'. Sammen danner de et unikt fingeraftryk.Forskningsdatasæt
-
Replication data from: What Comes Next? Evaluating Uncertainty in Neural Text Generators Against Human Production Variability
Giulianelli, M. (Ophavsmand), Baan, J. (Ophavsmand), Aziz, W. (Ophavsmand), Fernández, R. (Ophavsmand) & Plank, B. (Ophavsmand), ZENODO, 20 okt. 2023
DOI: 10.5281/zenodo.10025272, https://zenodo.org/records/10025272
Datasæt