Table of Contents
Fetching ...

Beyond identifiability: Learning causal representations with few environments and finite samples

Inbeom Lee, Tongtong Jin, Bryon Aragam

Abstract

We provide explicit, finite-sample guarantees for learning causal representations from data with a sublinear number of environments. Causal representation learning seeks to provide a rigourous foundation for the general representation learning problem by bridging causal models with latent factor models in order to learn interpretable representations with causal semantics. Despite a blossoming theory of identifiability in causal representation learning, estimation and finite-sample bounds are less well understood. We show that causal representations can be learned with only a logarithmic number of unknown, multi-node interventions, and that the intervention targets need not be carefully designed in advance. Through a careful perturbation analysis, we provide a new analysis of this problem that guarantees consistent recovery of (a) the latent causal graph, (b) the mixing matrix and representations, and (c) \emph{unknown} intervention targets.

Beyond identifiability: Learning causal representations with few environments and finite samples

Abstract

We provide explicit, finite-sample guarantees for learning causal representations from data with a sublinear number of environments. Causal representation learning seeks to provide a rigourous foundation for the general representation learning problem by bridging causal models with latent factor models in order to learn interpretable representations with causal semantics. Despite a blossoming theory of identifiability in causal representation learning, estimation and finite-sample bounds are less well understood. We show that causal representations can be learned with only a logarithmic number of unknown, multi-node interventions, and that the intervention targets need not be carefully designed in advance. Through a careful perturbation analysis, we provide a new analysis of this problem that guarantees consistent recovery of (a) the latent causal graph, (b) the mixing matrix and representations, and (c) \emph{unknown} intervention targets.

Paper Structure

This paper contains 21 sections, 10 theorems, 29 equations.

Key Result

Theorem 2.1

$\mathrm{(Identifiability ~ with ~ unknown~intervention~environments)}$ Under the model (eq.linSEM-eq.linmix) and assum.PSSS.0-assum.dnvr.0, the following can be identified from $K$ unknown, multi-node intervention environments:

Theorems & Definitions (13)

  • Theorem 2.1
  • Corollary 2.2
  • Lemma 3.1
  • Remark 3.1
  • Theorem 4.1: Informal
  • Lemma 4.2
  • Lemma 4.3
  • Lemma 4.4
  • Remark 4.1
  • Theorem 4.5
  • ...and 3 more