Table of Contents
Fetching ...

Contrastive Conformal Sets

Yahya Alkhatib, Wee Peng Tay

Abstract

Contrastive learning produces coherent semantic feature embeddings by encouraging positive samples to cluster closely while separating negative samples. However, existing contrastive learning methods lack principled guarantees on coverage within the semantic feature space. We extend conformal prediction to this setting by introducing minimum-volume covering sets equipped with learnable generalized multi-norm constraints. We propose a method that constructs conformal sets guaranteeing user-specified coverage of positive samples while maximizing negative sample exclusion. We establish theoretically that volume minimization serves as a proxy for negative exclusion, enabling our approach to operate effectively even when negative pairs are unavailable. The positive inclusion guarantee inherits the distribution-free coverage property of conformal prediction, while negative exclusion is maximized through learned set geometry optimized on a held-out training split. Experiments on simulated and real-world image datasets demonstrate improved inclusion-exclusion trade-offs compared to standard distance-based conformal baselines.

Contrastive Conformal Sets

Abstract

Contrastive learning produces coherent semantic feature embeddings by encouraging positive samples to cluster closely while separating negative samples. However, existing contrastive learning methods lack principled guarantees on coverage within the semantic feature space. We extend conformal prediction to this setting by introducing minimum-volume covering sets equipped with learnable generalized multi-norm constraints. We propose a method that constructs conformal sets guaranteeing user-specified coverage of positive samples while maximizing negative sample exclusion. We establish theoretically that volume minimization serves as a proxy for negative exclusion, enabling our approach to operate effectively even when negative pairs are unavailable. The positive inclusion guarantee inherits the distribution-free coverage property of conformal prediction, while negative exclusion is maximized through learned set geometry optimized on a held-out training split. Experiments on simulated and real-world image datasets demonstrate improved inclusion-exclusion trade-offs compared to standard distance-based conformal baselines.

Paper Structure

This paper contains 8 sections, 1 theorem, 6 equations, 1 figure.

Key Result

Corollary 1

If the positive inclusion condition in eq:pos-inc-alpha is satisfied by some positive generation mechanism $p_{\mathrm{pos}}$ at level $1-\alpha$, and the negative exclusion probability in eq:max-neg-exc is bounded below by $1-\beta$ for some $\beta \in (0,1)$, then the oracle mechanisms $p_{\mathrm respectively. Here, $Y_{\mathrm{pos}}$ and $Y_{\mathrm{neg}}$ are the labels of the positive and ne

Figures (1)

  • Figure \docprefix1: 3D UMAP projection of the semantic features of an STL10 anchor point and its positive and negative samples.

Theorems & Definitions (1)

  • Corollary 1: Oracle Inclusion-Exclusion Bounds