Table of Contents
Fetching ...

Robust Matrix Estimation with Side Information

Anish Agarwal, Jungjun Choi, Ming Yuan

Abstract

We introduce a flexible framework for high-dimensional matrix estimation to incorporate side information for both rows and columns. Existing approaches, such as inductive matrix completion, often impose restrictive structure-for example, an exact low-rank covariate interaction term, linear covariate effects, and limited ability to exploit components explained only by one side (row or column) or by neither-and frequently omit an explicit noise component. To address these limitations, we propose to decompose the underlying matrix as the sum of four complementary components: (possibly nonlinear) interaction between row and column characteristics; row characteristic-driven component, column characteristic-driven component, and residual low-rank structure unexplained by observed characteristics. By combining sieve-based projection with nuclear-norm penalization, each component can be estimated separately and these estimated components can then be aggregated to yield a final estimate. We derive convergence rates that highlight robustness across a range of model configurations depending on the informativeness of the side information. We further extend the method to partially observed matrices under both missing-at-random and missing-not-at-random mechanisms, including block-missing patterns motivated by causal panel data. Simulations and a real-data application to tobacco sales show that leveraging side information improves imputation accuracy and can enhance treatment-effect estimation relative to standard low-rank and spectral-based alternatives.

Robust Matrix Estimation with Side Information

Abstract

We introduce a flexible framework for high-dimensional matrix estimation to incorporate side information for both rows and columns. Existing approaches, such as inductive matrix completion, often impose restrictive structure-for example, an exact low-rank covariate interaction term, linear covariate effects, and limited ability to exploit components explained only by one side (row or column) or by neither-and frequently omit an explicit noise component. To address these limitations, we propose to decompose the underlying matrix as the sum of four complementary components: (possibly nonlinear) interaction between row and column characteristics; row characteristic-driven component, column characteristic-driven component, and residual low-rank structure unexplained by observed characteristics. By combining sieve-based projection with nuclear-norm penalization, each component can be estimated separately and these estimated components can then be aggregated to yield a final estimate. We derive convergence rates that highlight robustness across a range of model configurations depending on the informativeness of the side information. We further extend the method to partially observed matrices under both missing-at-random and missing-not-at-random mechanisms, including block-missing patterns motivated by causal panel data. Simulations and a real-data application to tobacco sales show that leveraging side information improves imputation accuracy and can enhance treatment-effect estimation relative to standard low-rank and spectral-based alternatives.

Paper Structure

This paper contains 34 sections, 10 theorems, 135 equations, 9 figures, 3 tables.

Key Result

Theorem 3.1

Suppose that Assumptions asp:noise--asp:moment hold. Then,

Figures (9)

  • Figure 1: Missing pattern in MNAR case
  • Figure 2: AMSE under different values of $\alpha_r$
  • Figure 3: $(AMSE_{\mathrm{other}} - AMSE_{\mathrm{our}})/AMSE_{\mathrm{our}}$ under different values of $\alpha_r$
  • Figure 4: Performance comparison in the MAR case
  • Figure 5: Performance comparison in the MNAR case
  • ...and 4 more figures

Theorems & Definitions (10)

  • Theorem 3.1: Convergence rate
  • Corollary 3.2
  • Theorem 4.1: Convergence rate for the MAR case
  • Theorem 4.2: Convergence rate for the MNAR case
  • Lemma C.1
  • Lemma C.2
  • Lemma C.3
  • Lemma C.4
  • Lemma C.5
  • Lemma C.6