Non-Exchangeable Conformal Language Generation with Nearest Neighbors

Dennis Thomas Ulmer, Chrysoula Zerva, André Martins

Research output: Conference Article in Proceeding or Book/Report chapterArticle in proceedingsResearchpeer-review

Abstract

Quantifying uncertainty in automatically generated text is important for letting humans check potential hallucinations and making systems more reliable. Conformal prediction is an attractive framework to provide predictions imbued with statistical guarantees, however, its application to text generation is challenging since any i.i.d. assumptions are not realistic. In this paper, we bridge this gap by leveraging recent results on *non-exchangeable* conformal prediction, which still ensures bounds on coverage. The result, *non-exchangeable conformal nucleus sampling*, is a novel extension of the conformal prediction framework to generation based on nearest neighbors. Our method can be used post-hoc for an arbitrary model without extra training and supplies token-level, calibrated prediction sets equipped with statistical guarantees. Experiments in machine translation and language modeling show encouraging results in generation quality. By also producing tighter prediction sets with good coverage, we thus give a more theoretically principled way to perform sampling with conformal guarantees.
Original languageEnglish
Title of host publicationFindings of the Association for Computational Linguistics: EACL 2024
EditorsYvette Graham, Matthew Purver
Number of pages20
VolumeEACL
PublisherAssociation for Computational Linguistics
Publication date17 Mar 2024
Edition2024
Pages1909-1929
Publication statusPublished - 17 Mar 2024

Keywords

  • Uncertainty Quantification
  • Text Generation
  • Conformal Prediction
  • Non-Exchangeable Conformal Nucleus Sampling
  • Statistical Guarantees

Fingerprint

Dive into the research topics of 'Non-Exchangeable Conformal Language Generation with Nearest Neighbors'. Together they form a unique fingerprint.

Cite this