Connectome-Constrained Spiking Neural Networks: Olfactory Classification Study
Moberg et al., 2026
Abstract
Artificial neural networks achieve remarkable performance through learned weight distributions over largely arbitrary topologies, yet biological neural circuits evolved under strict metabolic and behavioral constraints over millions of years. Whether this evolved wiring encodes computational priors that confer advantages in learning efficiency, generalization, or representational sparsity over randomly initialized architectures remains an open question. We present a controlled comparison between a Spiking Neural Network whose recurrent synaptic topology is constrained to the Drosophila melanogaster olfactory circuit as mapped by the FlyWire connectome (Shiu et al., 2024) and three baselines: a degree-matched randomly-wired SNN, a sparsity-matched MLP, and a fully-connected parameter-matched MLP. All models are trained on the DoOR olfactory receptor response dataset for multi-class odor identity classification using surrogate gradient descent. We hypothesize that biologically-derived synaptic topology encodes an inductive bias toward more efficient or stable olfactory representations than randomly-initialized architectures of equivalent capacity. We evaluate convergence rate, peak classification accuracy, spike sparsity, and training stability across five-fold cross-validation and multiple random seeds. All code, connectome extraction pipelines, and trained weights are released to support reproducibility and extension to other FlyWire circuit subgraphs.
Connectome-Constrained Spiking Neural Networks: Olfactory Classification Study
What this experiment asked
Does using the actual wiring diagram of a fruit fly's nose as the architecture of an AI network help it learn better than a randomly-wired network of the same size?
How it was tested
Four AI models were trained to identify odors from receptor response data. All models had roughly the same number of trainable parameters (~1.25M). The only difference between the two key models was whether their internal connections matched real fly brain anatomy or were randomly shuffled.
Dataset
DoOR — a database of olfactory receptor responses compiled from multiple published studies. 24 receptor measurements per odor, multi-class classification task.
Protocol
5-fold cross-validation × 3 random seeds = 15 independent runs per model, 60 total training runs. Results reported as mean ± standard deviation.
Plain Language Summary
The Four Models
ConnectomeSNN
Spiking network whose internal connections are fixed to match real fly brain anatomy (FlyWire connectome). The primary model — what the hypothesis is about.
SNNShuffledSNN
Identical spiking network but connections are randomly shuffled. Same number of connections per neuron, different wiring. The topology control.
SNNSparseMLP
Standard neural network with the same density of connections as the connectome models, but no spiking dynamics. Isolates the effect of sparsity from topology.
DenseMLP
Fully connected standard neural network. The conventional deep learning baseline. Matched in total parameter count to the other models.
Summary Results
All values: mean ± standard deviation across 15 runs. Test accuracy is proportion correct (0–1 scale, so 0.455 = 45.5% correct). Higher is better for accuracy; lower is better for stopped epoch (faster convergence).
| Model | Type | Test Accuracy | Stopped Epoch | Spike Sparsity | Parameters |
|---|---|---|---|---|---|
|
ConnectomeSNN Biological topology |
SNN |
45.5% ± 1.2% |
27.0 ± 8.8 |
80.2% ± 0.6% |
1,235,892 |
|
ShuffledSNN Random topology |
SNN |
45.4% ± 1.3% |
24.7 ± 7.1 |
80.3% ± 0.7% |
1,235,892 |
|
SparseMLP Sparse standard network |
MLP |
48.9% ± 1.1% |
19.3 ± 5.2 |
— | 1,235,892 |
|
DenseMLP Fully connected baseline |
MLP |
49.2% ± 0.9% |
18.7 ± 4.8 |
— | 1,235,180 |
ConnectomeSNN
ShuffledSNN
SparseMLP
DenseMLP
Run-by-Run Comparison: ConnectomeSNN vs ShuffledSNN
Each row is one matched training run (same data fold, same random seed). This table is the basis for the sign test reported in the paper.
| Run | Fold | Seed | Connectome Epochs |
Shuffled Epochs |
Connectome Accuracy |
Shuffled Accuracy |
Conn Trained Longer? |
|---|---|---|---|---|---|---|---|
| 1 | 0 | 7 | 34 | 26 | 46.7% | 46.6% | Yes |
| 2 | 0 | 1010 | 34 | 37 | 47.9% | 48.1% | No |
| 3 | 0 | 2013 | 23 | 28 | 46.3% | 47.0% | No |
| 4 | 1 | 7 | 26 | 18 | 45.3% | 44.5% | Yes |
| 5 | 1 | 1010 | 26 | 21 | 44.6% | 44.0% | Yes |
| 6 | 1 | 2013 | 18 | 20 | 47.1% | 47.5% | No |
| 7 | 2 | 7 | 28 | 20 | 46.4% | 45.5% | Yes |
| 8 | 2 | 1010 | 16 | 22 | 45.2% | 45.6% | No |
| 9 | 2 | 2013 | 23 | 23 | 46.1% | 45.6% | Tie |
| 10 | 3 | 7 | 41 | 30 | 46.6% | 45.8% | Yes |
| 11 | 3 | 1010 | 30 | 35 | 44.8% | 46.3% | No |
| 12 | 3 | 2013 | 44 | 24 | 46.8% | 45.8% | Yes |
| 13 | 4 | 7 | 19 | 24 | 45.9% | 45.8% | No |
| 14 | 4 | 1010 | 21 | 42 | 44.8% | 45.2% | No |
| 15 | 4 | 2013 | 17 | 20 | 43.0% | 43.6% | No |
Fold = which partition of odor classes was held out. Seed = random initialization number (7, 1010, or 2013). Same fold + seed = identical data split and starting conditions for both models.
Glossary
Connectome
A complete map of every neuron and synapse in a nervous system. The FlyWire connectome maps the fruit fly (Drosophila melanogaster) brain at single-synapse resolution.
Spiking Neural Network (SNN)
An AI model that mimics biological neurons — neurons only "fire" when their activation crosses a threshold, producing sparse, event-driven computation rather than continuous signals.
MLP (Multilayer Perceptron)
A conventional fully-connected neural network. The standard architecture used in most deep learning applications. No spiking dynamics, no biological constraints.
Surrogate gradient
A mathematical trick that allows spiking networks to be trained with backpropagation, which normally requires smooth differentiable functions. Spikes are binary (on/off), which backprop cannot handle directly.
Spike sparsity
The fraction of neurons that are silent (not firing) at any given moment. Higher sparsity = fewer neurons active. Biological Kenyon cells show ~90–95% sparsity; these models showed ~80%.
Early stopping
Training is halted when validation accuracy stops improving for 5 consecutive epochs. "Stopped epoch" is how many epochs the model trained before this triggered.
Cross-validation (5-fold)
The dataset is divided into 5 groups. Each group takes a turn as the test set while the other 4 train the model. Gives 5 independent accuracy estimates; reduces dependence on any one data split.
Sign test
A simple statistical test that asks: if there were truly no difference, how likely is it to see this many outcomes in one direction just by chance? p = 0.696 means 69.6% chance of seeing this result randomly — not significant.
DoOR dataset
Database of Odorant Responses — a compiled dataset of olfactory receptor neuron measurements from Drosophila melanogaster, aggregated from multiple published studies.
Degree-preserving shuffle
The connections in the ShuffledSNN were randomized, but each neuron kept the same number of inputs and outputs as in the real connectome. Only the specific wiring pattern changed, not the connection density per neuron.
Summary of results from Connectome-Constrained Spiking Neural Networks: Olfactory Classification Study by Moberg et al., 2026
The full research paper detailing our hypothesis, methodology, and interpretation is currently being edited. Follow MIRE or check back here for updates when the preprint is published.