Recognition: 3 theorem links
A Gentle Introduction to Conformal Prediction and Distribution-Free Uncertainty Quantification
Pith reviewed 2026-05-09 01:19 UTC · model claude-opus-4-7
The pith
Any pre-trained model can be wrapped into prediction sets with guaranteed finite-sample coverage, regardless of how the model was built or what the data distribution is.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The paper organizes a body of work around a single thesis: any black-box predictor, however badly trained, can be wrapped in a short post-hoc calibration step that produces prediction sets guaranteed to contain the truth with user-specified probability, in finite samples, without assumptions on the model or the data distribution. The wrapper requires only a held-out calibration set, a scalar score function s(x,y) measuring disagreement between an input and a candidate label, and an empirical quantile of those scores. The authors argue this recipe — split conformal prediction — is general enough to cover classification, quantile regression, Bayesian posteriors, outlier detection, segmentation
What carries the argument
The core object is the empirical quantile q̂ of conformal scores s(X_i, Y_i) on a held-out calibration set, taken at level ⌈(n+1)(1−α)⌉/n; the prediction set is {y : s(X_test, y) ≤ q̂}. The work is done by exchangeability: the test score is equally likely to land in any of the n+1 gaps of the sorted calibration scores, which forces marginal coverage ≥ 1−α with no further assumption. Every extension (covariate shift, drift, risk control, group balance) is a re-weighting or re-grouping of this same quantile.
If this is right
- Any deployed predictor — including a frozen neural network whose internals are unavailable — can be retrofitted with calibrated prediction sets using only a few hundred labeled holdout points and a few lines of code.
- Improvements in uncertainty quantification reduce, in practice, to designing better score functions for a given task rather than to proving new coverage theorems.
- Coverage guarantees extend cleanly past miscoverage to false-negative rate, false-discovery rate, IOU, and other bounded losses, by tuning a threshold on calibration data at a slightly conservative level.
- Under known covariate shift, reweighting calibration scores by the likelihood ratio restores exact finite-sample coverage; under unknown drift, weighted calibration with a rolling window degrades coverage only in proportion to a total-variation distance.
- Conditional coverage — the same guarantee for every subgroup or every input — is provably unattainable in general, so practitioners must check feature- and size-stratified coverage as a routine diagnostic rather than expecting marginal coverage to imply fairness across groups.
Where Pith is reading between the lines
- The framework's real cost is hidden in the choice of score function: validity is free, but informativeness (small sets) inherits all the failure modes of the underlying model, so conformal prediction launders calibration but not capability.
- Because the guarantee is marginal over the calibration draw, two practitioners running the same procedure on different held-out sets will see coverage that differs by several percentage points; reporting a single conformal interval without the Beta-distribution caveat overstates what was actually controlled.
- The risk-control extension via multiple testing on a parameter grid effectively recasts uncertainty quantification as a hypothesis-testing problem, which suggests power-versus-conservativeness tradeoffs from multiple-comparison theory will increasingly drive practical performance.
- The drift bound's dependence on total-variation distance, which is essentially never measurable in deployment, means the time-series guarantees are honest about being heuristic — the actual safety in production comes from short windows and fast recalibration, not from a theorem.
Load-bearing premise
The whole guarantee rests on the calibration data and the future test point being interchangeable — drawn from the same distribution in a way that does not care about order. When that fails (real distribution shift, time series, selection bias), the guarantee degrades, and the patches the paper offers require either knowing the shift or guessing its size.
What would settle it
Run the split-conformal recipe at α=0.1 on a fresh i.i.d. classification or regression task with n≈1000 calibration points, repeat over many random splits, and check that the empirical coverage histogram matches the Beta(n+1−⌊(n+1)α⌋, ⌊(n+1)α⌋) distribution centered at 1−α. A systematic shortfall below 1−α on i.i.d. data, larger than the Beta-Binomial fluctuations the paper tabulates, would falsify the central claim.
read the original abstract
Black-box machine learning models are now routinely used in high-risk settings, like medical diagnostics, which demand uncertainty quantification to avoid consequential model failures. Conformal prediction is a user-friendly paradigm for creating statistically rigorous uncertainty sets/intervals for the predictions of such models. Critically, the sets are valid in a distribution-free sense: they possess explicit, non-asymptotic guarantees even without distributional assumptions or model assumptions. One can use conformal prediction with any pre-trained model, such as a neural network, to produce sets that are guaranteed to contain the ground truth with a user-specified probability, such as 90%. It is easy-to-understand, easy-to-use, and general, applying naturally to problems arising in the fields of computer vision, natural language processing, deep reinforcement learning, and so on. This hands-on introduction is aimed to provide the reader a working understanding of conformal prediction and related distribution-free uncertainty quantification techniques with one self-contained document. We lead the reader through practical theory for and examples of conformal prediction and describe its extensions to complex machine learning tasks involving structured outputs, distribution shift, time-series, outliers, models that abstain, and more. Throughout, there are many explanatory illustrations, examples, and code samples in Python. With each code sample comes a Jupyter notebook implementing the method on a real-data example; the notebooks can be accessed and easily run using our codebase.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The manuscript is a self-contained tutorial on conformal prediction (CP) and related distribution-free uncertainty quantification techniques. It presents split CP with the standard marginal coverage guarantee 1−α ≤ P(Y∈C(X)) ≤ 1−α+1/(n+1) (Theorem 1, Appendix D), walks the reader through canonical score functions (APS §2.1, CQR §2.2, scalar uncertainty §2.3, Bayes-optimal §2.4), discusses adaptivity diagnostics and finite-sample coverage fluctuations (§3, Appendix C), and surveys extensions: group/class-conditional CP (§4.1–4.2), conformal risk control (§4.3), outlier detection (§4.4), covariate shift via weighting (§4.5), distribution drift (§4.6), full/cross/CV+ CP (§6), and Learn-then-Test for general risk control (Appendix A–B). Worked examples on Imagenet, MS-COCO, tumor segmentation, weather time-series, and toxic-comment detection are accompanied by short Python snippets and Jupyter notebooks. A historical section (§7) traces the development of CP from algorithmic randomness through to current trends.
Significance. The paper is explicitly expository and does not claim new theorems; its value lies in pedagogy, breadth of coverage, accurate attribution, and reproducibility. As an introduction it succeeds: the split-CP recipe is given in ~10 lines of NumPy with a correct ⌈(n+1)(1−α)⌉/n quantile correction, the proof in Appendix D is the standard exchangeability argument and is correctly stated, and the limits of the framework (marginal vs. conditional coverage, the impossibility result of [87], the continuity requirement for the upper bound, the unmeasurable TV distances in Theorem 4) are honestly disclosed at the relevant points (§3.1, §3.2, footnote 1, §5.3). The accompanying code and notebooks support reproducibility, and the bibliography is comprehensive and current. Tutorials of this scope and accuracy are genuinely useful to the community and have been heavily cited; the manuscript meets the standard for an accepted survey/tutorial.
minor comments (11)
- [§1, Eq. (1) and footnote 1] The upper bound 1−α+1/(n+1) requires continuous (tie-free) scores, as later stated in Theorem D.2 and footnote 1. Because Eq. (1) is the very first display in the paper and many readers will only skim, it would help to attach a short parenthetical at Eq. (1) itself (e.g., 'upper bound assumes continuous scores; see Thm. D.2') rather than deferring this caveat to a footnote and Appendix D. The current presentation could leave readers using discrete softmax outputs with the impression that they get the two-sided bound without randomized tie-breaking.
- [§1, calibration recipe] When introducing ˆq as the ⌈(n+1)(1−α)⌉/n empirical quantile, it would be worth pointing out explicitly that this requires α ≥ 1/(n+1); otherwise the algorithm returns the trivial set C(X)=Y. This corner case is handled implicitly in the proof of Theorem 1 but is not flagged in the main-text recipe, and beginners running the code with very small calibration sets may be confused.
- [§2.1, Eq. (3)] The +1 in 'k = sup{...} + 1' that ensures non-empty sets is stated without comment. A one-line note that this corresponds to the randomized correction of [4] omitted for simplicity (and a pointer to the linked notebook for the randomized version) would help reproducibility, since the deterministic version slightly over-covers.
- [§3.2, Table 1] Table 1 reports n(ε) for δ=0.1, α=0.1 only. Given that the surrounding text suggests n≈1000 as a rule of thumb, a second column or remark for at least one other (α,δ) pair (say α=0.05) would make the guidance more transferable; otherwise readers may misapply the n=1000 heuristic outside the regime in which it was computed.
- [§4.5, weighted CP] The display defining ˆq(x) as the 1−α quantile of a reweighted distribution silently assumes the scores have been pre-sorted (the manuscript notes this 'for notational convenience'). For a tutorial, an explicit version with general (unsorted) scores, or at least a sentence stating that ties and ordering require care, would prevent implementation bugs. The accompanying code does not appear in this section.
- [§4.6, Theorem 4] The bound 1−α−2Σw̃_iε_i contains TV distances ε_i that the manuscript itself acknowledges are 'never known' (§5.3). This is fine as stated, but readers would benefit from a sharper sentence next to Theorem 4 that the bound is best read as a structural statement (not a deployable certificate) and that the practical justification of the fixed-window/decay weights in §5.3 is heuristic. Currently this honest caveat is somewhat buried at the end of §5.3.
- [§5.5, Eq. (15) and ˆR+] The selective-classification example invokes the LTT machinery and a Binomial CDF upper bound, but the symbol δ is introduced only inside ˆR+(λ) without the user-facing reminder that the guarantee is now (1−δ) over the calibration draw rather than marginal. Given that earlier sections emphasized α as the only knob, a sentence flagging this shift from expectation-style to high-probability-style guarantees would aid the reader before Appendix A formalizes it.
- [Appendix A, Hoeffding p-value] The Hoeffding p-value in §A.1.1 assumes losses bounded in [0,1]. This is implicit in the surrounding text but not stated at the point of definition. Since the Appendix is meant to be a self-contained crash course, the boundedness assumption should be made explicit alongside the formula.
- [§7, Historical Notes] The historical section is engaging but in places mixes biographical anecdote with technical history in a way that is hard for a non-specialist to parse (e.g., the Bernoulli sequences/randomness deficiency thread). A short signposting sentence ('readers in a hurry can skip to Current Trends') would help, and the link from randomness deficiency to nonconformity scores deserves one more concrete sentence to make the connection clear.
- [Code listings (Figs. 2, 3, 5, 7, 12, 20, 23, 24)] The code samples mix np.quantile with method='higher' (Fig. 2) and interpolation='higher' (Figs. 3, 5) — the latter is the deprecated NumPy keyword. Standardizing on the current keyword and noting the NumPy version assumed would prevent silent deprecation warnings or errors for new users.
- [Figure 11 / Appendix C] The Beta(n+1−l, l) distribution of conditional coverage (with l=⌊(n+1)α⌋) is stated and plotted, but the exact relationship between this and the practical 'n≈1000' guideline could be tightened with one displayed inequality (e.g., a Hoeffding-style tail bound for the Beta) so the reader can compute n for their own (α, ε, δ) without running the notebook.
Simulated Author's Rebuttal
The referee recommends acceptance and raises no major comments, judging the tutorial accurate, honestly scoped, well-attributed, and reproducible. As there are no specific revision requests, we have nothing substantive to contest or amend. We thank the referee for the careful reading and for confirming that the central guarantees, the proof in Appendix D, the discussion of conditional-coverage limits, and the bibliography are correctly and fairly presented. We will use the opportunity of the next revision only for minor typographical polishing and to refresh pointers to recent follow-up work, leaving the technical content reviewed by the referee unchanged.
read point-by-point responses
-
Referee: The referee recommends acceptance and raises no major comments, noting that the tutorial succeeds at its pedagogical aims, that the split-CP recipe and Appendix D proof are correctly stated, that limitations (marginal vs. conditional coverage, impossibility of [87], continuity for the upper bound, unmeasurable TV distances in Theorem 4) are honestly disclosed at the relevant points, and that the accompanying code/notebooks and bibliography are comprehensive and current.
Authors: We thank the referee for the careful and thorough reading of the manuscript and for the positive recommendation. We are grateful that the referee has verified the correctness of the central technical statements (Theorem 1 and the Appendix D proof, the ⌈(n+1)(1−α)⌉/n quantile correction in the code), the breadth and currency of the references, and the explicit disclosure of the framework's limitations at the relevant points (footnote 1 on tie-breaking, §3.1 on marginal vs. conditional coverage, §3.2 and Appendix C on finite-sample fluctuations, and §5.3 on the unmeasurability of the TV distances appearing in Theorem 4). Since the referee raised no major comments, no substantive revisions are required in response to this report. We will, however, take the opportunity of the next arXiv revision to fix any minor typographical issues that have been brought to our attention by readers since the v6 posting, and to refresh pointers to fast-moving follow-up work (e.g., online/adaptive conformal under distribution shift and conformal risk control), without altering the technical content the referee has reviewed. We thank the referee again for engaging with the manuscript in detail. revision: no
Circularity Check
No circularity: tutorial reproducing standard, externally-attributed results with a textbook proof.
full rationale
This paper is an expository introduction to conformal prediction. Its central technical claim — Theorem 1 / Theorem D.1, the split-conformal marginal coverage guarantee 1−α ≤ P(Y_test ∈ C(X_test)) ≤ 1−α+1/(n+1) — is attributed to Vovk, Gammerman, and Saunders [5] and proved in Appendix D by the standard exchangeability-of-ranks argument: under exchangeability of (s_1,...,s_n,s_test), P(s_test ≤ s_(k)) = k/(n+1), which immediately yields the bound when ˆq is set to s_⌈(n+1)(1−α)⌉. The proof's hypotheses (exchangeability, quantile definition) do not contain the conclusion; the conclusion follows from a combinatorial fact about ranks of exchangeable variables, which is independent of the present authors. Other major results are similarly attributed and proved or cited externally: CQR (Theorem implied, citing Romano et al. [8]), conformal risk control (Theorem 2, citing [17]), weighted/covariate-shift conformal (Theorem 3, citing Tibshirani et al. [25]), drift (Theorem 4, citing Barber et al. [26]), full conformal (Theorem 5, citing [1]), and Learn-then-Test (Theorem A.1, citing [18]). Self-citations exist (e.g., [4], [17], [18] include the present authors), but they are not load-bearing for the central marginal-coverage theorem, which predates the authors. No "prediction" in the paper is a fitted quantity renamed; ˆq is explicitly a calibration-set quantile and the coverage statement is a probabilistic statement about a held-out test point. The paper is honest about scope limitations the skeptic raised (marginal vs. conditional coverage in §3.1 citing impossibility result [87]; tie-breaking for the upper bound in footnote 1 and §6.1; unmeasurable TV distances in Theorem 4 acknowledged in §5.3). None of these are circular reductions; they are disclosed assumptions. The derivation chain is self-contained against external mathematical facts (exchangeability, order statistics, Hoeffding/Bentkus concentration, Bonferroni/sequential testing), and the cited results are mathematically standard and verifiable. Score: 0.
Axiom & Free-Parameter Ledger
Forward citations
Cited by 35 Pith papers
-
An Optimal Sauer Lemma Over $k$-ary Alphabets
A sharp Sauer inequality for multiclass and list prediction is established in terms of the DS dimension, tight for every alphabet size k, list size ℓ, and dimension value.
-
Adaptive Stopping for Multi-Turn LLM Reasoning
MiCP is the first conformal prediction method for multi-turn LLM pipelines that allocates per-turn error budgets to enable adaptive stopping with an overall coverage guarantee, shown to reduce turns and cost on RAG an...
-
GRAPHLCP: Structure-Aware Localized Conformal Prediction on Graphs
GRAPHLCP improves localized conformal prediction on graphs by using feature-aware densification and Personalized PageRank kernels to incorporate topology for better coverage and efficiency.
-
TRACE: Transport Alignment Conformal Prediction via Diffusion and Flow Matching Models
TRACE creates valid conformal prediction sets for complex generative models by scoring outputs via averaged denoising or velocity errors along stochastic transport paths instead of likelihoods.
-
When Does Trimming Help Conformal Prediction? A Retained-Law Diagnostic under Calibration Contamination
Trimming helps conformal prediction under contamination precisely when the anomaly score separates retention probabilities without biasing clean scores, otherwise the retained mixture coefficient prevents substantial ...
-
In-Context Positive-Unlabeled Learning
PUICL is a transformer pretrained on synthetic PU data from structural causal models that solves positive-unlabeled classification via in-context learning without gradient updates or fitting.
-
Delving into Non-Exchangeability for Conformal Prediction in Graph-Structured Multivariate Time Series
SCALE uses Spectral Graph Conditional Exchangeability (SGCE) and graph wavelets to achieve valid coverage and improved efficiency in conformal prediction for non-exchangeable graph time series by conformalizing high-f...
-
SURE-RAG: Sufficiency and Uncertainty-Aware Evidence Verification for Selective Retrieval-Augmented Generation
SURE-RAG aggregates pair-level claim-evidence relations into interpretable signals for selective RAG answering, reaching 0.9075 Macro-F1 on HotpotQA-RAG v3 while providing auditability and reducing unsafe answers by 3...
-
Intrinsic effective sample size for manifold-valued Markov chain Monte Carlo via kernel discrepancy
An intrinsic effective sample size for manifold MCMC is defined via kernel discrepancy as the number of independent draws yielding equivalent expected squared discrepancy to the target.
-
Profile Likelihood Inference for Anisotropic Hyperbolic Wrapped Normal Models on Hyperbolic Space
The profile maximum likelihood estimator for the location in anisotropic hyperbolic wrapped normal models is strongly consistent, asymptotically normal, and attains the Hájek-Le Cam minimax lower bound under squared g...
-
Query-Efficient Quantum Approximate Optimization via Graph-Conditioned Trust Regions
A GNN predicts Gaussians over QAOA parameters to create graph-conditioned trust regions that reduce circuit evaluations for MaxCut from 85-343 down to 45 while keeping approximation ratios within 3 points of heuristics.
-
Adaptive Conformal Anomaly Detection with Time Series Foundation Models for Signal Monitoring
A model-agnostic adaptive conformal anomaly detection approach uses weighted quantile bounds learned from past foundation model predictions to deliver interpretable p-value scores with stable calibration under shifts ...
-
Causal inference for social network formation
Random team assignments in a professional firm reveal that indirect ties strongly increase new direct tie formation, while effects of degree and local density are smaller and less robust.
-
Diagnosing LLM Judge Reliability: Conformal Prediction Sets and Transitivity Violations
LLM judges display per-document transitivity violations in 33-67% of cases despite low aggregate rates, while conformal prediction set widths serve as reliable indicators of document-level difficulty with cross-judge ...
-
Conformal Margin Risk Minimization: An Envelope Framework for Robust Learning under Label Noise
CMRM adds a conformal quantile regularization on prediction margins to any loss, improving noisy-label classification accuracy up to 3.39% across methods and benchmarks while preserving performance at zero noise.
-
Conformal Risk Control under Non-Monotone Losses: Theory and Finite-Sample Guarantees
Conformal risk control for bounded non-monotone losses over a grid of size m achieves excess risk of order sqrt(log m / n) with n calibration samples, which is minimax optimal.
-
Multi-Fidelity Quantile Regression
A model-agnostic two-stage estimator links high-fidelity quantiles to low-fidelity ones via a covariate-dependent level function for faster convergence and better accuracy with limited high-fidelity data.
-
CONTRA: Conformal Prediction Region via Normalizing Flow Transformation
CONTRA generates sharp multi-dimensional conformal prediction regions by defining nonconformity scores as distances from the center in the latent space of a normalizing flow.
-
Scale selection for geometric medians on product manifolds
Joint location-scale minimization for geometric medians on product manifolds degenerates to marginal medians, and three new scale-selection methods restore identifiability with asymptotic guarantees.
-
Conformal Agent Error Attribution
A new filtration-based conformal prediction method attributes errors in multi-agent systems by producing contiguous sequence sets with finite-sample coverage guarantees, enabling rollback recovery.
-
Networked Information Aggregation for Binary Classification
Sequential prediction passing on DAGs for logistic regression yields O(M/sqrt(D)) excess loss when M-agent windows cover all features, with Omega(k/D) lower bound identifying depth as the fundamental limit.
-
Unsupervised Confidence Calibration for Reasoning LLMs from a Single Generation
Unsupervised single-generation confidence calibration for reasoning LLMs via offline self-consistency proxy distillation outperforms baselines on math and QA tasks and improves selective prediction.
-
DAG-STL: A Hierarchical Framework for Zero-Shot Trajectory Planning under Signal Temporal Logic Specifications
DAG-STL decomposes long-horizon STL planning into decomposition, timed waypoint allocation, and diffusion-based trajectory generation to enable zero-shot planning under unknown dynamics.
-
Answer Only as Precisely as Justified: Calibrated Claim-Level Specificity Control for Agentic Systems
Compositional selective specificity (CSS) improves overcommitment-aware utility from 0.846 to 0.913 on LongFact while retaining 0.938 specificity by calibrating claim-level backoffs in agentic AI responses.
-
Blind-Spot Mass: A Good-Turing Framework for Quantifying Deployment Coverage Risk in Machine Learning Systems
Blind-spot mass uses Good-Turing unseen-species estimation to measure the total probability of states with low empirical support, showing that 95% of operational mass lies in blind spots at tau=5 across wearable activ...
-
Adaptive Conformal Prediction for Reliable and Explainable Medical Image Classification
An adaptive lambda criterion for RAPS achieves 95.72% global coverage and at least 90% coverage across all difficulty strata on medical image datasets while keeping average prediction set size at 1.09.
-
Quantile-Free Uncertainty Quantification in Graph Neural Networks
QpiGNN provides a quantile-free dual-head architecture for GNN uncertainty quantification that directly optimizes coverage and interval width, yielding 22% higher coverage and 50% narrower intervals than baselines on ...
-
Towards Dependable Retrieval-Augmented Generation Using Factual Confidence Prediction
A conformal prediction filter for retrieval chunks plus an attention-based factuality classifier can raise RAG answer quality by up to 6% and detect inconsistent generations up to 77% of the time.
-
An empirical evaluation of the risks of AI model updates using clinical data: stability, arbitrariness, and fairness
Updating clinical AI models can cause prediction flips, arbitrariness, and unfair error rates across groups, requiring dedicated monitoring dimensions.
-
ReconVLA: An Uncertainty-Guided and Failure-Aware Vision-Language-Action Framework for Robotic Control
ReconVLA enhances pretrained vision-language-action robotic policies with conformal prediction for uncertainty estimation and failure detection without retraining.
-
Uncertainty-Aware Transformers: Conformal Prediction for Language Models
CONFIDE applies conformal prediction to transformer embeddings for valid prediction sets, improving accuracy up to 4.09% and efficiency over baselines on models like BERT-tiny.
-
Probably Approximately Correct (PAC) Guarantees for Data-Driven Reachability Analysis: A Theoretical and Empirical Comparison
Formal connections between PAC bounds for three data-driven reachability methods are established, with empirical results showing they are not interchangeable despite similarities.
-
Neural posterior estimation for scalable and accurate inverse parameter inference in Li-ion batteries
NPE delivers millisecond-scale parameter inference for Li-ion batteries that matches or exceeds Bayesian calibration accuracy while adding local sensitivity interpretability, though with higher voltage prediction errors.
-
AIVV: Neuro-Symbolic LLM Agent-Integrated Verification and Validation for Trustworthy Autonomous Systems
AIVV deploys LLM agents in a council to semantically validate anomalies in time-series data against natural-language requirements, automating human-in-the-loop verification for autonomous systems.
-
Uncertainty in Physics and AI: Taxonomy, Quantification, and Validation
A unified taxonomy of uncertainty in ML for physics is introduced together with validation tools such as coverage, calibration, and proper scoring rules, illustrated on regression and classification tasks.
Reference graph
Works this paper leans on
-
[1]
V. Vovk, A. Gammerman, and G. Shafer, Algorithmic Learning in a Random World . Springer, 2005
work page 2005
-
[2]
Inductive confidence machines for regression,
H. Papadopoulos, K. Proedrou, V. Vovk, and A. Gammerman, “Inductive confidence machines for regression,” in Machine Learning: European Conference on Machine Learning , 2002, pp. 345–356
work page 2002
-
[3]
Distribution-free prediction bands for non-parametric regression,
J. Lei and L. Wasserman, “Distribution-free prediction bands for non-parametric regression,” Journal of the Royal Statistical Society: Series B: Statistical Methodology , pp. 71–96, 2014
work page 2014
-
[4]
Uncertainty sets for image classifiers using conformal prediction,
A. N. Angelopoulos, S. Bates, J. Malik, and M. I. Jordan, “Uncertainty sets for image classifiers using conformal prediction,” in International Conference on Learning Representations, 2021
work page 2021
-
[5]
Machine-learning applications of algorithmic random- ness,
V. Vovk, A. Gammerman, and C. Saunders, “Machine-learning applications of algorithmic random- ness,” in International Conference on Machine Learning , 1999, pp. 444–453
work page 1999
-
[6]
Least ambiguous set-valued classifiers with bounded error levels,
M. Sadinle, J. Lei, and L. Wasserman, “Least ambiguous set-valued classifiers with bounded error levels,” Journal of the American Statistical Association , vol. 114, pp. 223–234, 2019
work page 2019
-
[7]
Mauricio Sadinle, Jing Lei, and Larry Wasserman
Y. Romano, M. Sesia, and E. J. Cand` es, “Classification with valid and adaptive coverage,”arXiv:2006.02544, 2020
-
[8]
Conformalized quantile regression,
Y. Romano, E. Patterson, and E. Cand` es, “Conformalized quantile regression,” inAdvances in Neural Information Processing Systems, vol. 32, 2019, pp. 3543–3553
work page 2019
-
[9]
R. Koenker and G. Bassett Jr, “Regression quantiles,” Econometrica: Journal of the Econometric Society, vol. 46, no. 1, pp. 33–50, 1978
work page 1978
-
[10]
A. N. Angelopoulos, A. P. Kohli, S. Bates, M. I. Jordan, J. Malik, T. Alshaabi, S. Upadhyayula, and Y. Romano, “Image-to-image regression with distribution-free uncertainty quantification and applications in imaging,” arXiv preprint arXiv:2202.05265 , 2022. 32
-
[11]
Bayes-optimal prediction with frequentist coverage control,
P. Hoff, “Bayes-optimal prediction with frequentist coverage control,” arXiv:2105.14045, 2021
-
[12]
L. Wasserman, “Frasian inference,” Statistical Science, vol. 26, no. 3, pp. 322–325, 2011
work page 2011
-
[13]
Comparing the bayes and typicalness frame- works,
T. Melluish, C. Saunders, I. Nouretdinov, and V. Vovk, “Comparing the bayes and typicalness frame- works,” in European Conference on Machine Learning, Springer, 2001, pp. 360–371
work page 2001
-
[14]
Conditional validity of inductive conformal predictors,
V. Vovk, “Conditional validity of inductive conformal predictors,” in Proceedings of the Asian Con- ference on Machine Learning, vol. 25, 2012, pp. 475–490
work page 2012
-
[15]
Knowing what you know: Valid and validated confidence sets in multiclass and multilabel prediction,
M. Cauchois, S. Gupta, and J. Duchi, “Knowing what you know: Valid and validated confidence sets in multiclass and multilabel prediction,” arXiv:2004.10181, 2020
-
[16]
Improving conditional coverage via orthogonal quantile regression,
S. Feldman, S. Bates, and Y. Romano, “Improving conditional coverage via orthogonal quantile regression,” in Advances in Neural Information Processing Systems , 2021
work page 2021
-
[17]
arXiv preprint arXiv:2208.02814 , year=
A. N. Angelopoulos, S. Bates, A. Fisch, L. Lei, and T. Schuster, “Conformal risk control,” arXiv preprint arXiv:2208.02814, 2022
-
[18]
A. N. Angelopoulos, S. Bates, E. J. Cand` es, M. I. Jordan, and L. Lei, “Learn then test: Calibrating predictive algorithms to achieve risk control,” arXiv:2110.01052, 2021
-
[19]
A review of novelty detection,
M. A. Pimentel, D. A. Clifton, L. Clifton, and L. Tarassenko, “A review of novelty detection,” Signal Processing, vol. 99, pp. 215–249, 2014
work page 2014
-
[20]
R. A. Fisher, “Design of experiments,” British Medical Journal, vol. 1, no. 3923, p. 554, 1936
work page 1936
-
[21]
Significance tests which may be applied to samples from any populations,
E. J. Pitman, “Significance tests which may be applied to samples from any populations,” Supplement to the Journal of the Royal Statistical Society , vol. 4, no. 1, pp. 119–130, 1937
work page 1937
-
[22]
Testing exchangeability on-line,
V. Vovk, I. Nouretdinov, and A. Gammerman, “Testing exchangeability on-line,” in Proceedings of the 20th International Conference on Machine Learning (ICML-03) , 2003, pp. 768–775
work page 2003
-
[23]
Prediction and outlier detection in classification problems,
L. Guan and R. Tibshirani, “Prediction and outlier detection in classification problems,” arXiv:1905.04396, 2019
-
[24]
Testing for outliers with conformal p-values,
S. Bates, E. Cand` es, L. Lei, Y. Romano, and M. Sesia, “Testing for outliers with conformal p-values,” arXiv:2104.08279, 2021
-
[25]
Conformal prediction under covariate shift,
R. J. Tibshirani, R. Foygel Barber, E. Candes, and A. Ramdas, “Conformal prediction under covariate shift,” in Advances in Neural Information Processing Systems 32 , 2019, pp. 2530–2540
work page 2019
-
[26]
Candes, Aaditya Ramdas, and Ryan J
R. F. Barber, E. J. Candes, A. Ramdas, and R. J. Tibshirani, “Conformal prediction beyond ex- changeability,” arXiv:2202.13415, 2022
-
[27]
Conformal prediction with localization.arXiv preprint arXiv:1908.08558,
L. Guan, “Conformal prediction with localization,” arXiv:1908.08558, 2020
-
[28]
Microsoft coco: Common objects in context,
T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Doll´ ar, and C. L. Zitnick, “Microsoft coco: Common objects in context,” in European conference on computer vision, Springer, 2014, pp. 740–755
work page 2014
-
[29]
A. Malinin, N. Band, G. Chesnokov, Y. Gal, M. J. Gales, A. Noskov, A. Ploskonosov, L. Prokhorenkova, I. Provilkov, V. Raina, et al., “Shifts: A dataset of real distributional shift across multiple large-scale tasks,” arXiv preprint arXiv:2107.07455 , 2021
-
[30]
CatBoost: gradient boosting with categorical features support
A. V. Dorogush, V. Ershov, and A. Gulin, “Catboost: Gradient boosting with categorical features support,” arXiv preprint arXiv:1810.11363 , 2018
work page Pith review arXiv 2018
-
[31]
Adaptive Conformal Inference Under Distribution Shift , year =
I. Gibbs and E. Cand` es, “Adaptive conformal inference under distribution shift,” arXiv:2106.00170, 2021
-
[32]
Adaptive conformal predictions for time series,
M. Zaffran, O. F´ eron, Y. Goude, J. Josse, and A. Dieuleveut, “Adaptive conformal predictions for time series,” in International Conference on Machine Learning , PMLR, 2022, pp. 25 834–25 866
work page 2022
-
[33]
Conformal inference for online prediction with arbitrary distribution shifts, 2023
I. Gibbs and E. Cand` es, “Conformal inference for online prediction with arbitrary distribution shifts,” arXiv preprint arXiv:2208.08401 , 2022
-
[34]
Conformal prediction interval for dynamic time-series,
C. Xu and Y. Xie, “Conformal prediction interval for dynamic time-series,” in International Confer- ence on Machine Learning , PMLR, 2021, pp. 11 559–11 569. 33
work page 2021
-
[35]
Hanu and Unitary team, Detoxify, Github
L. Hanu and Unitary team, Detoxify, Github. https://github.com/unitaryai/detoxify, 2020
work page 2020
-
[36]
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “Bert: Pre-training of deep bidirectional trans- formers for language understanding,” arXiv preprint arXiv:1810.04805 , 2018
work page Pith review arXiv 2018
-
[37]
Wilds: A benchmark of in-the-wild distribution shifts,
P. W. Koh, S. Sagawa, H. Marklund, S. M. Xie, M. Zhang, A. Balsubramani, W. Hu, M. Yasunaga, R. L. Phillips, I. Gao, et al., “Wilds: A benchmark of in-the-wild distribution shifts,” in International Conference on Machine Learning, PMLR, 2021, pp. 5637–5664
work page 2021
-
[38]
A tutorial on conformal prediction,
G. Shafer and V. Vovk, “A tutorial on conformal prediction,” Journal of Machine Learning Research, vol. 9, no. Mar, pp. 371–421, 2008
work page 2008
-
[39]
Computing full conformal prediction set with approximate homotopy,
E. Ndiaye and I. Takeuchi, “Computing full conformal prediction set with approximate homotopy,” in Advances in Neural Information Processing Systems , 2019
work page 2019
-
[40]
Root-finding approaches for computing conformal prediction set,
E. Ndiaye and I. Takeuchi, “Root-finding approaches for computing conformal prediction set,” Ma- chine Learning, 2022
work page 2022
-
[41]
V. Vovk, “Cross-conformal predictors,” Annals of Mathematics and Artificial Intelligence , vol. 74, no. 1-2, pp. 9–28, 2015
work page 2015
-
[42]
Predictive inference with the jack- knife+,
R. F. Barber, E. J. Candes, A. Ramdas, and R. J. Tibshirani, “Predictive inference with the jack- knife+,” The Annals of Statistics , vol. 49, no. 1, pp. 486–507, 2021
work page 2021
-
[43]
Exact and asymptotically robust permutation tests,
E. Chung and J. P. Romano, “Exact and asymptotically robust permutation tests,” The Annals of Statistics, vol. 41, no. 2, pp. 484–507, 2013
work page 2013
-
[44]
On a test of whether one of two random variables is stochastically larger than the other,
H. B. Mann and D. R. Whitney, “On a test of whether one of two random variables is stochastically larger than the other,” The Annals of Mathematical Statistics , pp. 50–60, 1947
work page 1947
-
[45]
E. L. Lehmann, “The power of rank tests,” The Annals of Mathematical Statistics , pp. 23–43, 1953
work page 1953
- [46]
-
[47]
B. Efron and R. J. Tibshirani, An introduction to the bootstrap. CRC press, 1994
work page 1994
-
[48]
Distribution-free cumulative sum control charts using bootstrap-based control limits,
S. Chatterjee and P. Qiu, “Distribution-free cumulative sum control charts using bootstrap-based control limits,” The Annals of Applied Statistics , vol. 3, no. 1, pp. 349–369, 2009
work page 2009
-
[49]
G. T. Fechner, Kollektivmasslehre. Engelmann, 1897
-
[50]
Grundlagen der wahrscheinlichkeitsrechnung,
R. von Mises, “Grundlagen der wahrscheinlichkeitsrechnung,” Mathematische Zeitschrift, vol. 5, no. 1, pp. 52–99, 1919
work page 1919
-
[51]
Die widerspruchfreiheit des kollectivbegriffes der wahrscheinlichkeitsrechnung,
A. Wald, “Die widerspruchfreiheit des kollectivbegriffes der wahrscheinlichkeitsrechnung,” Ergebnisse Eines Mathematischen Kolloquiums , vol. 8, no. 38-72, p. 37, 1937
work page 1937
-
[52]
On the concept of a random sequence,
A. Church, “On the concept of a random sequence,” Bulletin of the American Mathematical Society , vol. 46, no. 2, pp. 130–135, 1940
work page 1940
-
[53]
Etude critique de la notion de collectif,
J. Ville, “Etude critique de la notion de collectif,” Bull. Amer. Math. Soc , vol. 45, no. 11, p. 824, 1939
work page 1939
-
[54]
The sources of Kolmogorov’s Grundbegriffe,
G. Shafer and V. Vovk, “The sources of Kolmogorov’s Grundbegriffe,” Statistical Science, vol. 21, no. 1, pp. 70–98, 2006
work page 2006
-
[55]
Kolmogorov’s complexity conception of probability,
V. Vovk, “Kolmogorov’s complexity conception of probability,” Synthese Library, pp. 51–70, 2001
work page 2001
-
[56]
Kolmogorov on the role of randomness in probability theory,
C. P. Porter, “Kolmogorov on the role of randomness in probability theory,” Mathematical Structures in Computer Science , vol. 24, no. 3, 2014
work page 2014
-
[57]
Three approaches to the quantitative definition of information,
A. N. Kolmogorov, “Three approaches to the quantitative definition of information,” Problems of Information Transmission, vol. 1, no. 1, pp. 1–7, 1965
work page 1965
-
[58]
Logical basis for information theory and probability theory,
A. Kolmogorov, “Logical basis for information theory and probability theory,” IEEE Transactions on Information Theory , vol. 14, no. 5, pp. 662–664, 1968
work page 1968
-
[59]
Combinatorial foundations of information theory and the calculus of probabili- ties,
A. N. Kolmogorov, “Combinatorial foundations of information theory and the calculus of probabili- ties,” Russian Mathematical Surveys , vol. 38, no. 4, pp. 29–40, 1983. 34
work page 1983
-
[60]
On the concept of the Bernoulli property,
V. G. Vovk, “On the concept of the Bernoulli property,” Russian Mathematical Surveys, vol. 41, no. 1, p. 247, 1986
work page 1986
-
[61]
V. Vovk, “Testing randomness online,” Statistical Science, vol. 36, no. 4, pp. 595–611, 2021
work page 2021
-
[62]
Sophistication as randomness deficiency,
F. Mota, S. Aaronson, L. Antunes, and A. Souto, “Sophistication as randomness deficiency,” in International Workshop on Descriptional Complexity of Formal Systems , Springer, 2013, pp. 172– 181
work page 2013
-
[63]
Determination of sample sizes for setting tolerance limits,
S. S. Wilks, “Determination of sample sizes for setting tolerance limits,” Annals of Mathematical Statistics, vol. 12, no. 1, pp. 91–96, 1941
work page 1941
-
[64]
Statistical prediction with special reference to the problem of tolerance limits,
——, “Statistical prediction with special reference to the problem of tolerance limits,” Annals of Mathematical Statistics, vol. 13, no. 4, pp. 400–409, 1942
work page 1942
-
[65]
An extension of Wilks’ method for setting tolerance limits,
A. Wald, “An extension of Wilks’ method for setting tolerance limits,” Annals of Mathematical Statis- tics, vol. 14, no. 1, pp. 45–55, 1943
work page 1943
-
[66]
J. W. Tukey, “Non-parametric estimation II. Statistically equivalent blocks and tolerance regions–the continuous case,” Annals of Mathematical Statistics , vol. 18, no. 4, pp. 529–539, 1947
work page 1947
-
[67]
Finite exchangeable sequences,
P. Diaconis and D. Freedman, “Finite exchangeable sequences,” The Annals of Probability , pp. 745– 764, 1980
work page 1980
-
[68]
Exchangeability and related topics,
D. J. Aldous, “Exchangeability and related topics,” in ´Ecole d’ ´Et´ e de Probabilit´ es de Saint-Flour XIII—1983, 1985, pp. 1–198
work page 1983
-
[69]
Funzione caratteristica di un fenomeno aleatorio,
B. De Finetti, “Funzione caratteristica di un fenomeno aleatorio,” in Atti del Congresso Internazionale dei Matematici: Bologna del 3 al 10 de Settembre di 1928 , 1929, pp. 179–190
work page 1928
-
[70]
D. A. Freedman, “Bernard Friedman’s urn,” The Annals of Mathematical Statistics , pp. 956–970, 1965
work page 1965
-
[71]
Symmetric measures on Cartesian products,
E. Hewitt and L. J. Savage, “Symmetric measures on Cartesian products,” Transactions of the Amer- ican Mathematical Society, vol. 80, no. 2, pp. 470–501, 1955
work page 1955
-
[72]
J. F. Kingman, “Uses of exchangeability,” The Annals of Probability, vol. 6, no. 2, pp. 183–197, 1978
work page 1978
-
[73]
Dobriban, Topics in Modern Statistical Learning (STAT 991, UPenn, 2022 Spring) , Dec
E. Dobriban, Topics in Modern Statistical Learning (STAT 991, UPenn, 2022 Spring) , Dec. 2022
work page 2022
-
[74]
A. Gammerman, V. Vovk, and V. Vapnik, “Learning by transduction,” Proceedings of the Fourteenth Conference on Uncertainty in Artificial Intelligence , vol. 14, pp. 148–155, 1998
work page 1998
-
[75]
Transduction with confidence and credibility,
C. Saunders, A. Gammerman, and V. Vovk, “Transduction with confidence and credibility,” 1999
work page 1999
-
[76]
On-line confidence machines are well-calibrated,
V. Vovk, “On-line confidence machines are well-calibrated,” in The 43rd Annual IEEE Symposium on Foundations of Computer Science , IEEE, 2002, pp. 187–196
work page 2002
-
[77]
Self-calibrating probability forecasting.,
V. Vovk, G. Shafer, and I. Nouretdinov, “Self-calibrating probability forecasting.,” in Neural Infor- mation Processing Systems, 2003, pp. 1133–1140
work page 2003
-
[78]
Venn-abers predictors.arXiv preprint arXiv:1211.0025,
V. Vovk and I. Petej, “Venn-Abers predictors,” arXiv:1211.0025, 2012
-
[79]
Nonparametric predictive distributions based on conformal prediction,
V. Vovk, J. Shen, V. Manokhin, and M.-g. Xie, “Nonparametric predictive distributions based on conformal prediction,” Machine Learning, pp. 1–30, 2017
work page 2017
-
[80]
Efficient nonparametric conformal prediction regions,
J. Lei, J. Robins, and L. Wasserman, “Efficient nonparametric conformal prediction regions,” arXiv:1111.1418, 2011
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.