Table of Contents
Fetching ...

Computing with Living Neurons: Chaos-Controlled Reservoir Computing with Knowledge Transplant

Seung Hyun Kim, Zhi Dou, Gaurav Upadhyay, Anay Pattanaik, Leo Maslov, Lav Varshney, John Beggs, Howard Gritton, Mattia Gazzola

Abstract

We introduce chaos-controlled Reservoir Computing (cc-RC) for living neural cultures: dynamically rich substrates of unique potential for adaptive computation. To account for intrinsic biological variability, cc-RC combines: (i) pre-training identification of each culture's dynamical signature and phase-portrait attractor; (ii) low-power optical chaos control to stabilize spontaneous and stimulus-evoked activity; (iii) readout training within this controlled regime. Across hundreds of neural samples, cc-RC enables robust learning and pattern classification, improving both accuracy and model longevity by approximately 300% over standard RC. We further propose Knowledge Transplant (KT), for which the reservoir map learned by an expert culture is transplanted to an attractor-equivalent student culture, reducing training time to minutes while improving performance. By enabling cross-substrate, reusable learned models, KT paves the way for knowledge accumulation and sharing across neural populations, transcending biological lifespan limits.

Computing with Living Neurons: Chaos-Controlled Reservoir Computing with Knowledge Transplant

Abstract

We introduce chaos-controlled Reservoir Computing (cc-RC) for living neural cultures: dynamically rich substrates of unique potential for adaptive computation. To account for intrinsic biological variability, cc-RC combines: (i) pre-training identification of each culture's dynamical signature and phase-portrait attractor; (ii) low-power optical chaos control to stabilize spontaneous and stimulus-evoked activity; (iii) readout training within this controlled regime. Across hundreds of neural samples, cc-RC enables robust learning and pattern classification, improving both accuracy and model longevity by approximately 300% over standard RC. We further propose Knowledge Transplant (KT), for which the reservoir map learned by an expert culture is transplanted to an attractor-equivalent student culture, reducing training time to minutes while improving performance. By enabling cross-substrate, reusable learned models, KT paves the way for knowledge accumulation and sharing across neural populations, transcending biological lifespan limits.

Paper Structure

This paper contains 10 sections, 5 figures.

Figures (5)

  • Figure 1: Overview of cc-RC setup. (a) ChR2-transfected mESCs are expanded, formed into embryoid body (EB) in suspension culture, and differentiated into motoneurons. On day 7, differentiated EBs are dissociated and plated on the MEA at a density of $5,000\,\text{cells per }\mathrm{mm}^2$. Cultures mature by day 14. See SI for details. (b) Living neural networks plated on $128$-electrodes MEAs realize integrated neural chips. (c) MiV platforms enable the recording of electric local field potentials (LFPs) via PCB-MEA interfaces connected to Intan-RHS2116 chips. Optical inputs are delivered by a $473\,\mathrm{nm}$ laser (Doric). MiV platform and hosted cultures sit in an incubator for environmental control (temperature, humidity, $\mathrm{CO}_2$) and Faraday shielding. Acquisition system is placed outside to reduce degradation and signal noise. (d) LFPs show rich spatiotemporal dynamics: global avalanches, local bursts, and individual spikes. Neural activities are interpreted through state-space embedding to identify attractors and transitions, enabling cc-RC and comparison across preparations/conditions. (e) Neural cultures serve as a reservoirs driven by temporally encoded light pulses. Evoked firing rates are measured and processed through a readout map (linear vector) for task-specific operations.
  • Figure 2: Pre-flight characterization of reservoir type and quality. (a) Representative raster plots for spontaneous activity classes. Type A: Poisson-like incoherent activity. Type B/C: partially synchronized activity with individual spikes and global bursts, separated by burst rate (B: $<0.5\,\mathrm{Hz}$, C: $>0.5\,\mathrm{Hz}$). Type D: strongly synchronized activity with narrow, regular bursts and minimal inter-burst spiking. (b) Representative latent trajectories for each type. Type A shows compact and aperiodic dynamics. Type B/C span larger latent volumes with variable-period structure. Type D displays highly uniform laminar trajectories. (c) Representative streamline fields from latent-state velocity estimates. (d) Diagnostics across $500+$ spontaneous recordings: Type occurrences, branching ratio, kernel rank, spectral radius, and connectivity correlation. Red bands indicate preferred operating ranges per metric. Avalanche criticality analysis is not reported for Type A due to absent coherent bursts. Each data point is determined using at least $5\,\mathrm{min}$ of spontaneous activity. Details on metrics and experimental measurements can be found in the SI.
  • Figure 3: Performance of (naive) reservoir computing. (a) Ten temporally encoded input patterns. Pattern N contains N light pulses ($15\,\mathrm{ms}$ each) within $900\,\mathrm{ms}$, followed by $100\,\mathrm{ms}$ rest ($1$s total pattern window). Firing rates are measured at the end of each pattern window. (b) Training protocol. During $1$ hour, patterns are presented in random order and multichannel firing rates are recorded. A ridge-classification readout $W_\text{out}$ is fit with $10$-fold cross-validation. (c) Test protocol. After $15$ minutes rest, each round presents patterns over $3$ minutes ($180$ patterns), followed by $12$ minutes rest. The fixed readout $W_\text{out}$ is evaluated over $12$ hours ($47$ rounds). (d) Accuracy by culture Type (A--D): first hour (left) and full $12$-hour test (right). Accuracy declines over time, after the first $1$--$2$ hours. Black line indicates best performing samples (95th percentile). Insets show early and late latent trajectories corresponding to pattern 1, 2, 3: trajectories are initially well-separates, then collapse into overlapping clusters.
  • Figure 4: Chaos-controlled reservoir computing. (a) Chaos control strategy: a low-amplitude triangular modulation ($1~\mathrm{Hz}$, $100\%$ duty cycle, $10\%$ of input intensity) is superimposed during training and testing. Modulation starts $40$ seconds before pattern onset to suppress initial transients. (b) Burst-timing regularization. Raster plots without (top) and with (bottom) modulation; orange dots indicate burst onset. Modulation aligns spontaneous bursting with input patterns, and entrains burst-interval variance within $\pm1.5\,\mathrm{ms}$ over $60$ minutes of training. (c) Representative latent trajectories without and with modulation (no pattern input). Modulation improves phase alignment and trajectory consistency. (d) cc-RC accuracy over $12$ hours for Type C and D. Stabilized dynamics preserve separable pattern trajectories and extend useful performance, yielding $350\%$ longer time above the $60\%$ accuracy threshold, and $250\%$ higher accuracy at $5$ hours.
  • Figure 5: Knowledge Transplant. (a) KT workflow. An expert reservoir is fully trained to obtain $W_\text{out}^\text{e}$. Alignment via transformation $\mathcal{T}$ allows to transplant $W_\text{out}^\text{e}$ to a student sample $W_\text{out}^\text{s}=\mathcal{T}^{-1}W_\text{out}^\text{e}$ (b) Attractors' alignment. Student's attractor in latent space (estimated via few input-evoked trajectories) is mapped onto the expert's attractor (estimated via thousands of trajectories) via ridge-regularized regression to obtain the geometric transformation $\mathcal{T}$. (c) Learning curves from-scratch versus KT. For both approaches, accuracy at each time point is computed using a $70\%$--$30\%$ train–test split. For KT, the readout is updated over time using only the training partition, with evaluation on the held‑out test set, ensuring fair comparison. Score values normalized to final from-scratch accuracy: * $p<0.05$, ** $p<0.01$, *** $p<0.001$. (d) Representative KT cases. (top) KT sustained learning and drift-resistance vs. from-scratch learning where performance degrades after $15$ minutes. (bottom) KT bootstrapping and performance ceiling enhancement vs. from-scratch learning.