Abstract
In Natural Language Generation (NLG) tasks,
for any input, multiple communicative goals are
plausible, and any goal can be put into words,
or produced, in multiple ways. We characterise
the extent to which human production varies
lexically, syntactically, and semantically across
four NLG tasks, connecting human production
variability to aleatoric or data uncertainty. We
then inspect the space of output strings shaped
by a generation system’s predicted probability
distribution and decoding algorithm to probe
its uncertainty. For each test input, we measure
the generator’s calibration to human production
variability. Following this instance-level ap-
proach, we analyse NLG models and decoding
strategies, demonstrating that probing a genera-
tor with multiple samples and, when possible,
multiple references, provides the level of detail
necessary to gain understanding of a model’s
representation of uncertainty.
for any input, multiple communicative goals are
plausible, and any goal can be put into words,
or produced, in multiple ways. We characterise
the extent to which human production varies
lexically, syntactically, and semantically across
four NLG tasks, connecting human production
variability to aleatoric or data uncertainty. We
then inspect the space of output strings shaped
by a generation system’s predicted probability
distribution and decoding algorithm to probe
its uncertainty. For each test input, we measure
the generator’s calibration to human production
variability. Following this instance-level ap-
proach, we analyse NLG models and decoding
strategies, demonstrating that probing a genera-
tor with multiple samples and, when possible,
multiple references, provides the level of detail
necessary to gain understanding of a model’s
representation of uncertainty.
Original language | English |
---|---|
Title of host publication | Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing |
Number of pages | 22 |
Place of Publication | Singapore |
Publisher | Association for Computational Linguistics |
Publication date | Dec 2023 |
Pages | 14349–14371 |
DOIs | |
Publication status | Published - Dec 2023 |
Keywords
- Natural Language Generation
- Communicative goals
- Human production variability
- Aleatoric uncertainty
- Decoding algorithm
- Lexical variability
- Syntactic variability
- Semantic variability
- Model calibration
- Uncertainty representation
Fingerprint
Dive into the research topics of 'What Comes Next? Evaluating Uncertainty in Neural Text Generators Against Human Production Variability'. Together they form a unique fingerprint.Datasets
-
Replication data from: What Comes Next? Evaluating Uncertainty in Neural Text Generators Against Human Production Variability
Giulianelli, M. (Creator), Baan, J. (Creator), Aziz, W. (Creator), Fernández, R. (Creator) & Plank, B. (Creator), ZENODO, 20 Oct 2023
DOI: 10.5281/zenodo.10025272, https://zenodo.org/records/10025272
Dataset