Statistical Error Bounds for GANs with Nonlinear Objective Functionals
Jeremiah Birrell
TL;DR
This work addresses statistical guarantees for GANs built on nonlinear objective functionals via the $(f,\Gamma)$-GAN framework, which unifies $f$-divergences with a discriminator space to produce nonlinear objectives. It introduces the generalized cumulant generating function $\Lambda_f^P$ and derives finite-sample concentration inequalities for the GAN estimator using novel Rademacher-complexity analyses, without requiring compact support. A key contribution is an error decomposition that separates optimization, discriminator-approximation, and statistical errors, with explicit bounds that depend on terms such as $\mathcal{R}_{\widetilde{\Gamma},Q,n}$ and $\Delta_{f,m}$. An additional result provides finite-second-moment-based bounds for Rademacher complexities under unbounded support, enabling applicability to heavy-tailed data and neural-networks, and linking to the IPM-GAN case in the linear-$f^*$ limit.
Abstract
Generative adversarial networks (GANs) are unsupervised learning methods for training a generator distribution to produce samples that approximate those drawn from a target distribution. Many such methods can be formulated as minimization of a metric or divergence between probability distributions. Recent works have derived statistical error bounds for GANs that are based on integral probability metrics (IPMs), e.g., WGAN which is based on the 1-Wasserstein metric. In general, IPMs are defined by optimizing a linear functional (difference of expectations) over a space of discriminators. A much larger class of GANs, which we here call $(f,Γ)$-GANs, can be constructed using $f$-divergences (e.g., Jensen-Shannon, KL, or $α$-divergences) together with a regularizing discriminator space $Γ$ (e.g., $1$-Lipschitz functions). These GANs have nonlinear objective functions, depending on the choice of $f$, and have been shown to exhibit improved performance in a number of applications. In this work we derive statistical error bounds for $(f,Γ)$-GANs for general classes of $f$ and $Γ$ in the form of finite-sample concentration inequalities. These results prove the statistical consistency of $(f,Γ)$-GANs and reduce to the known results for IPM-GANs in the appropriate limit. Our results use novel Rademacher complexity bounds which provide new insight into the performance of IPM-GANs for distributions with unbounded support and have application to statistical learning tasks beyond GANs.
