Table of Contents
Fetching ...

Spectral methods: crucial for machine learning, natural for quantum computers?

Vasilis Belis, Joseph Bowles, Rishabh Gupta, Evan Peters, Maria Schuld

Abstract

This article presents an argument for why quantum computers could unlock new methods for machine learning. We argue that spectral methods, in particular those that learn, regularise, or otherwise manipulate the Fourier spectrum of a machine learning model, are often natural for quantum computers. For example, if a generative machine learning model is represented by a quantum state, the Quantum Fourier Transform allows us to manipulate the Fourier spectrum of the state using the entire toolbox of quantum routines, an operation that is usually prohibitive for classical models. At the same time, spectral methods are surprisingly fundamental to machine learning: A spectral bias has recently been hypothesised to be the core principle behind the success of deep learning; support vector machines have been known for decades to regularise in Fourier space, and convolutional neural nets build filters in the Fourier space of images. Could, then, quantum computing open fundamentally different, much more direct and resource-efficient ways to design the spectral properties of a model? We discuss this potential in detail here, hoping to stimulate a direction in quantum machine learning research that puts the question of ``why quantum?'' first.

Spectral methods: crucial for machine learning, natural for quantum computers?

Abstract

This article presents an argument for why quantum computers could unlock new methods for machine learning. We argue that spectral methods, in particular those that learn, regularise, or otherwise manipulate the Fourier spectrum of a machine learning model, are often natural for quantum computers. For example, if a generative machine learning model is represented by a quantum state, the Quantum Fourier Transform allows us to manipulate the Fourier spectrum of the state using the entire toolbox of quantum routines, an operation that is usually prohibitive for classical models. At the same time, spectral methods are surprisingly fundamental to machine learning: A spectral bias has recently been hypothesised to be the core principle behind the success of deep learning; support vector machines have been known for decades to regularise in Fourier space, and convolutional neural nets build filters in the Fourier space of images. Could, then, quantum computing open fundamentally different, much more direct and resource-efficient ways to design the spectral properties of a model? We discuss this potential in detail here, hoping to stimulate a direction in quantum machine learning research that puts the question of ``why quantum?'' first.

Paper Structure

This paper contains 32 sections, 1 theorem, 74 equations, 8 figures.

Key Result

Theorem 1

Let $p(x)=|\langle x | \psi_{\theta}\rangle|^2$ be the measurement distribution of a (trainable) quantum state, with Fourier transform $\hat{p}(k)$. Let $\hat{\psi}(k)$ be the Fourier coefficients of the amplitudes $\psi(x) = \langle x | \psi_{\theta}\rangle$. Then

Figures (8)

  • Figure 1: A common simplicity bias of machine learning models is their smoothness, which is linked to a decay of the model function's Fourier spectrum. But designing such a decay in Fourier space is usually computationally expensive. Can quantum computers help here?
  • Figure 2: A sketch of the idea of smoothing an empirical distribution of training samples in Fourier space. The empirical distribution is sparse, with support only on the training data. Therefore, its Fourier spectrum is dense and has support on high-order frequencies which can be seen as a consequence of finite data effects. By applying a low-pass filter in Fourier space we impose smoothness on the resulting distribution.
  • Figure 3: An example of smoothing an empirical distribution in Fourier space to solve the generative learning problem. Starting from an empirical distribution $p_{\mathcal{X}}$ of four binary data samples (top left), we compute the (Walsh) Fourier spectrum (bottom left). We then apply a filter of the form $\hat{p}_{\theta}(k) = (1-2\theta)^{|k|} \hat{p}_{\mathcal{X}}(k)$, where $\theta \in [0,1]$ controls the spectral decay rate. Moving back into direct space, we get a model that generalised from the training data by making the empirical distribution smoother.
  • Figure 4: An example of smoothing empirical distributions with quantum computers, using the same filter as in Figure \ref{['fig:smoothing_classical']}. Fourier filtering is performed on the amplitudes of the quantum state. We encode the empirical distribution (top left) into a superposition $\ket{\psi_{\mathcal{X}}}$ (mid left), to which we apply the (Walsh) Fourier transform (bottom left). We then apply a non-unitary transformation (with the help of ancillas) to implement the filter $\hat{\psi}_{\theta}(k) \propto (1-2\theta)^{|k|} \hat{\psi}_{\mathcal{X}}(k)$ (bottom right). Moving back into direct space (mid right), we get a quantum model with a new distribution of amplitudes, whose measurement distribution constitutes the generative model. Comparing to Figure \ref{['fig:smoothing_classical']}, it is clear that manipulating the Fourier coefficients of the amplitudes of a quantum model leads to a different generative model, but with a similar structure: high probabilities are amplified and low ones suppressed.
  • Figure 5: Illustration of the relation between a group and a homogeneous space, which are spaces where every point can be reached by acting with a group element. Group elements that map to the same point live in cosets of the stabiliser subgroup (whose elements map $x_0$ to itself). We can associate the homogeneous space with the quotient space $G/H$.
  • ...and 3 more figures

Theorems & Definitions (7)

  • Definition 1
  • Definition 2
  • Definition 3
  • Definition 4
  • Definition 5
  • Theorem 1
  • proof