Coding schemes in neural networks learning classification tasks
Alexander van Meegen, Haim Sompolinsky
TL;DR
The paper addresses how neural networks learn interpretable, task-dependent representations by analyzing the Bayesian weight posterior in the non-lazy, mean-field regime and how the observed coding schemes depend on neuronal nonlinearity. It develops a theory that yields closed-form single-neuron posteriors and posterior-averaged kernels, revealing three distinct coding schemes—analog (linear), redundant (sigmoidal), and sparse (ReLU)—that emerge as the hidden nonlinearity changes. Across toy tasks and standard datasets like MNIST and CIFAR10, the work shows that non-lazy networks learn strong features but improve generalization mainly through reduced predictor variance rather than mean-predictor improvements, with symmetry breaking and neural collapse concepts offering a unifying view of the representations. The study highlights how weight-scaling and nonlinearity profoundly shape representations, offering insights into representation learning, generalization, and potential transfer learning implications in deep networks, while noting regime-specific limitations and the role of regularization in the observed phenomena.
Abstract
Neural networks posses the crucial ability to generate meaningful representations of task-dependent features. Indeed, with appropriate scaling, supervised learning in neural networks can result in strong, task-dependent feature learning. However, the nature of the emergent representations, which we call the `coding scheme', is still unclear. To understand the emergent coding scheme, we investigate fully-connected, wide neural networks learning classification tasks using the Bayesian framework where learning shapes the posterior distribution of the network weights. Consistent with previous findings, our analysis of the feature learning regime (also known as `non-lazy', `rich', or `mean-field' regime) shows that the networks acquire strong, data-dependent features. Surprisingly, the nature of the internal representations depends crucially on the neuronal nonlinearity. In linear networks, an analog coding scheme of the task emerges. Despite the strong representations, the mean predictor is identical to the lazy case. In nonlinear networks, spontaneous symmetry breaking leads to either redundant or sparse coding schemes. Our findings highlight how network properties such as scaling of weights and neuronal nonlinearity can profoundly influence the emergent representations.
