Trade-off between Gradient Measurement Efficiency and Expressivity in Deep Quantum Neural Networks
Koki Chinzei, Shinichiro Yamano, Quoc Hoan Tran, Yasuhiro Endo, Hirotaka Oshima
TL;DR
This work establishes a fundamental trade-off between gradient measurement efficiency and expressivity in deep quantum neural networks, showing that higher expressivity (larger dynamical Lie algebras) necessitates greater gradient measurement costs. The authors formalize gradient measurability via a dynamical Lie algebra graph and derive two key inequalities: $\mathcal{X}_{\rm exp} \leq \frac{4^n}{\mathcal{F}_{\rm eff}} - \mathcal{F}_{\rm eff}$ and $\mathcal{X}_{\rm exp} \geq \mathcal{F}_{\rm eff}$, highlighting a fundamental limit on efficient gradient estimation. To approach this limit, they introduce the stabilizer-logical product ansatz (SLPA), a commuting-block circuit built from stabilizers and logical Pauli operators that saturates the trade-off upper bound by exploiting symmetry; SLPA enables gradient estimation with only $2B$ (or $2B-1$) measurement types per block, greatly reducing sample complexity in practice. Numerical experiments on symmetric-function learning and quantum phase recognition demonstrate that SLPA achieves high accuracy and trainability while dramatically lowering data requirements compared with parameter-shift-based approaches, underscoring the practical impact of symmetry-aware, efficient gradient estimation for variational quantum algorithms.
Abstract
Quantum neural networks (QNNs) require an efficient training algorithm to achieve practical quantum advantages. A promising approach is gradient-based optimization, where gradients are estimated by quantum measurements. However, QNNs currently lack general quantum algorithms for efficiently measuring gradients, which limits their scalability. To elucidate the fundamental limits and potentials of efficient gradient estimation, we rigorously prove a trade-off between gradient measurement efficiency (the mean number of simultaneously measurable gradient components) and expressivity in deep QNNs. This trade-off indicates that more expressive QNNs require higher measurement costs per parameter for gradient estimation, while reducing QNN expressivity to suit a given task can increase gradient measurement efficiency. We further propose a general QNN ansatz called the stabilizer-logical product ansatz (SLPA), which achieves the trade-off upper bound by exploiting the symmetric structure of the quantum circuit. Numerical experiments show that the SLPA drastically reduces the sample complexity needed for training while maintaining accuracy and trainability compared to well-designed circuits based on the parameter-shift method.
