AlignedCut: Visual Concepts Discovery on Brain-Guided Universal Feature Space
Huzheng Yang, James Gee, Jianbo Shi
TL;DR
AlignedCut introduces a brain-guided universal channel alignment that maps layer-wise deep-net features from multiple models into a shared space using brain voxel fMRI as supervision. By coupling a linear channel transform with brain prediction, the method discovers recurring visual concepts via spectral clustering, revealing figure-ground representations before category semantics and enabling cross-model concept tracing without a decoder. The approach shows concept emergence across layers, consistent brain ROI mappings, and interpretable layer-to-layer dynamics, aided by a scalable Nystrom-like spectral approximation and eigen-constraints. This yields a principled framework to quantify how visual information flows through networks and how shared concepts organize across models, with potential impact on model interpretability and cross-domain alignment.
Abstract
We study the intriguing connection between visual data, deep networks, and the brain. Our method creates a universal channel alignment by using brain voxel fMRI response prediction as the training objective. We discover that deep networks, trained with different objectives, share common feature channels across various models. These channels can be clustered into recurring sets, corresponding to distinct brain regions, indicating the formation of visual concepts. Tracing the clusters of channel responses onto the images, we see semantically meaningful object segments emerge, even without any supervised decoder. Furthermore, the universal feature alignment and the clustering of channels produce a picture and quantification of how visual information is processed through the different network layers, which produces precise comparisons between the networks.
