arXiv Analytics

Sign in

arXiv:2402.10978 [cs.LG]AbstractReferencesReviewsResources

Language Models with Conformal Factuality Guarantees

Christopher Mohri, Tatsunori Hashimoto

Published 2024-02-15Version 1

Guaranteeing the correctness and factuality of language model (LM) outputs is a major open problem. In this work, we propose conformal factuality, a framework that can ensure high probability correctness guarantees for LMs by connecting language modeling and conformal prediction. We observe that the correctness of an LM output is equivalent to an uncertainty quantification problem, where the uncertainty sets are defined as the entailment set of an LM's output. Using this connection, we show that conformal prediction in language models corresponds to a back-off algorithm that provides high probability correctness guarantees by progressively making LM outputs less specific (and expanding the associated uncertainty sets). This approach applies to any black-box LM and requires very few human-annotated samples. Evaluations of our approach on closed book QA (FActScore, NaturalQuestions) and reasoning tasks (MATH) show that our approach can provide 80-90% correctness guarantees while retaining the majority of the LM's original output.

Related articles: Most relevant | Search more
arXiv:2411.06590 [cs.LG] (Published 2024-11-10)
CriticAL: Critic Automation with Language Models
arXiv:2403.19839 [cs.LG] (Published 2024-03-28)
The New Agronomists: Language Models are Experts in Crop Management
arXiv:2404.03683 [cs.LG] (Published 2024-04-01)
Stream of Search (SoS): Learning to Search in Language