Pattern theory: the mathematics of perception
David Mumford
TL;DR
Pattern Theory advances a unifying mathematical framework for perception rooted in Bayesian inference and graphical models, extending from simple HMMs to rich continuous and geometric representations. It shows how MRFs, BBP, and continuum models capture grouping, segmentation, and shape in sensory data, and surveys computational tools such as particle filtering, variational methods, and diffusion-based approaches. The work highlights challenges from non-Markov dependencies and heavy-tailed statistics, while connecting statistical perception to geometric shape analysis via diffeomorphisms and geodesic flows. Overall, it argues for a scalable, pattern-driven paradigm that could enable fully unsupervised learning and robust perception across speech and vision domains.
Abstract
Is there a mathematical theory underlying intelligence? Control theory addresses the output side, motor control, but the work of the last 30 years has made clear that perception is a matter of Bayesian statistical inference, based on stochastic models of the signals delivered by our senses and the structures in the world producing them. We will start by sketching the simplest such model, the hidden Markov model for speech, and then go on illustrate the complications, mathematical issues and challenges that this has led to.
