Unsupervised Conformal Inference: Bootstrapping and Alignment to Control LLM Uncertainty

📅 2025-09-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Quantifying generation uncertainty in black-box large language models (LLMs) is challenging due to the absence of token-level probabilities and ground-truth labels. Method: We propose the first unsupervised conformal inference framework for LLMs, leveraging the geometric structure of response embeddings to construct a Gram-matrix-based atypicality score. It integrates bootstrap-enhanced uncertainty calibration (UCP) and conformal alignment to enable user-predicate-driven, statistically rigorous threshold calibration—without accessing model internals or ground-truth labels. Contribution/Results: Relying solely on response embeddings and bootstrap-residual aggregation, our framework achieves uncertainty calibration and hallucination filtering. Experiments across multiple benchmarks show near-nominal coverage, significantly reduced hallucination rates, tighter and more stable thresholds, and superior performance over lightweight detectors—while incurring comparable computational overhead.

Technology Category

Application Category

📝 Abstract
Deploying black-box LLMs requires managing uncertainty in the absence of token-level probability or true labels. We propose introducing an unsupervised conformal inference framework for generation, which integrates: generative models, incorporating: (i) an LLM-compatible atypical score derived from response-embedding Gram matrix, (ii) UCP combined with a bootstrapping variant (BB-UCP) that aggregates residuals to refine quantile precision while maintaining distribution-free, finite-sample coverage, and (iii) conformal alignment, which calibrates a single strictness parameter $τ$ so a user predicate (e.g., factuality lift) holds on unseen batches with probability $ge 1-α$. Across different benchmark datasets, our gates achieve close-to-nominal coverage and provide tighter, more stable thresholds than split UCP, while consistently reducing the severity of hallucination, outperforming lightweight per-response detectors with similar computational demands. The result is a label-free, API-compatible gate for test-time filtering that turns geometric signals into calibrated, goal-aligned decisions.
Problem

Research questions and friction points this paper is trying to address.

Managing uncertainty in black-box LLMs without token probabilities or labels
Developing unsupervised conformal inference for generation with distribution-free coverage
Calibrating alignment parameters to reduce hallucination severity in LLMs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Unsupervised conformal inference framework for generation
Bootstrapping variant aggregates residuals for quantile precision
Conformal alignment calibrates parameter for user predicates
L
Lingyou Pang
Department of Statistics, University of California, Davis
L
Lei Huang
Department of Statistics, University of California, Davis
Jianyu Lin
Jianyu Lin
Sr. Machine Learning Engineer @ Intuitive Surgical, Inc.
Computer VisionMachine LearningMedical Image Analysis
T
Tianyu Wang
Department of Applied Mathematics and Statistics, Johns Hopkins University
A
Akira Horiguchi
Department of Statistics, University of California, Davis
A
Alexander Aue
Department of Statistics, University of California, Davis
Carey E. Priebe
Carey E. Priebe
Professor of Applied Mathematics and Statistics, Johns Hopkins University
statistical inference for high-dimensional and graph data