- Pascal's Chatbot Q&As
- Posts
- The structure of convolutional architectures alone is sufficient to produce representations aligned with primate visual cortex, even without training.
The structure of convolutional architectures alone is sufficient to produce representations aligned with primate visual cortex, even without training.
This challenges the longstanding belief that cortex-like representations arise from extensive learning on natural visual input.
Cortex-Aligned Visual Representations Emerging “de novo” in Convolutional Architectures
by ChatGPT-5.1
The bioRxiv preprint and the Nature Machine Intelligence article present a striking and provocative thesis: many of the hallmark representational properties of primate and human visual cortex emerge spontaneously in wide convolutional neural networks even without pretraining, without exposure to natural images, and with minimal or no synaptic tuning. This finding significantly upends commonly held assumptions in both computational neuroscience and AI development, where cortical alignment has typically been attributed to large-scale supervised or self-supervised training on massive visual datasets.
The authors demonstrate that two architectural inductive biases — (1) spatial compression via pooling, and (2) feature expansion via increased channel dimensionality— are sufficient to generate cortex-aligned representational geometry across multiple stages of visual processing, including comparisons to monkey electrophysiology and human fMRI data.
Their conclusion is profound: CNN architecture alone, under specific dimensionality transformations, is close enough to the biological visual system that cortical-like representations emerge before learning.
This essay organizes the insights from the papers into (I) surprising statements, (II) controversial claims, (III) valuable findings for AI development, (IV) implications for future AI systems, and (V) an evaluation of how likely these insights will yield near-term benefits.
I. Surprising Statements and Findings
1. Cortex-aligned representations appear without training
The most surprising finding is that untrained convolutional networks — with random weights — approximate visual cortex representations. The authors show that these networks predict neural data from monkeys and humans at levels comparable to or above trained non-convolutional baselines.
This contradicts the widely held view that alignment arises primarily from exposure to natural image statistics and optimization signals.
2. Feature expansion dramatically improves alignment — but only in CNNs
Dimensionality expansion (increasing channels or “width”) enhances alignment, but this benefit vanishes when CNN inductive biases are removed or perturbed.
Thus: “Width matters only under convolutional architectural constraints.”
3. Spatial compression (pooling) is equally essential
The paper emphasizes that both spatial compression and feature expansion are jointly necessary; CNNs lacking either component performed poorly.
This indicates a synergistic effect rather than two independent contributing factors.
4. Lesion studies reveal that slight architectural disruptions eliminate cortical alignment
Targeted removal of convolutional inductive biases breaks the cortex-like geometry, implying the alignment is structural, not incidental.
5. Wide untrained CNNs outperform other architectures (Transformers, MLPs, hybrids)
Despite the enormous success of Transformers in vision, these models do not naturally exhibit cortical alignment before training.
This suggests that cortical alignment is not a universal property of neural networks, but specific to convolutional inductive biases.
6. Biological plausibility emerges from architectural constraints, not learning
This supports the idea that some aspects of cortical visual system organization — like Gabor-like filters, curvature tuning, and hierarchical invariances — may be consequences of architecture rather than experience.
II. Controversial Claims
1. Pretraining may be far less important for cortical alignment than believed
The prevailing view (supported by literature since Yamins & DiCarlo 2014) is that training on natural images produces cortex-like representations.
These papers challenge that assumption, showing training may mostly refine or amplify architectural predispositions rather than create them.
This is a major challenge to dominant narratives in neuroscience-aligned AI research.
2. Architectural inductive bias may explain more cortical properties than goal-driven optimization
If correct, this shifts the research frontier: instead of “brains and networks converge because of shared optimization objectives,” the explanation may be “brains and networks share structural constraints.”
This implicitly critiques the emphasis on task-driven models as mechanistic explanations of the brain.
3. Current brain–model alignment benchmarks may overestimate the role of training
Because alignment emerges in untrained CNNs, some brain-score results may reflect architectural similarity rather than genuine functional correspondence.
This could ignite debate over evaluation methods in computational neuroscience.
4. Transformers’ lack of native cortical alignment challenges their biological legitimacy
Vision Transformers have become dominant in high-performing AI systems; these results argue they diverge sharply from primate visual representations unless heavily trained.
Many researchers will dispute whether biological alignment should be a design criterion for AI at all.
III. Valuable Findings for AI Development
1. Convolution’s core inductive biases remain uniquely powerful
Despite years of attention to Transformer architectures, the evidence shows:
locality
weight sharing
hierarchical spatial compression
feature expansion
are crucial for biological alignment and may still be essential for efficient, generalizable vision models.
This challenges the notion that convolution is obsolete.
2. Training-free or low-training models may be more viable than expected
If wide CNNs already approximate cortical representations, training may merely refine representations rather than build them from scratch.
Possible applications:
energy-efficient pretraining
rapid specialization for edge devices
safer models with predictable inductive biases
3. Dimensionality engineering (compress/expand) offers a systematic design principle
Instead of ad hoc architecture tuning, designers can leverage the paper’s analytical framework to optimize network geometry for perceptual tasks.
4. Neuroscience-inspired architecture search can be simplified
Instead of emulating complex biological circuitry, the findings suggest the cortex may use simple geometric manipulations of dimensionality.
This provides a tractable architectural template for future models.
5. Implications for interpretability and safety
If cortical alignment arises structurally:
interpretability may improve because structure → function becomes predictable
safety mechanisms can be embedded architecturally rather than learned
models may exhibit more stable representational geometry across initializations
This reduces emergent unpredictability in training dynamics.
IV. How These Findings Could Impact Future AI Systems
1. Renewed interest in hybrid CNN–Transformer architectures
Many state-of-the-art vision models already incorporate convolutions (e.g., ConvNeXt, CoAtNet).
These results justify stronger convolutional backbones for cortical alignment and sample efficiency.
2. “Training-optional” architectures for embodied or robotic perception
Robotics will benefit from architectures that work well with:
minimal data
domain shift
unstructured sensory input
CNNs with built-in cortical alignment may outperform Transformers in low-data sensorimotor scenarios.
3. Improved artificial general intelligence for multimodal perception
If biological visual cortex and CNNs share deep structural commonalities, integrating CNN-like inductive biases into multimodal LLMs may improve:
grounding
scene understanding
invariance
robustness
4. Enhanced brain–computer interfaces and neural decoding
The paper uses correlations between untrained CNN features and neural recordings.
This could support more accurate:
decoding models
prosthetics
neural data compression
fMRI reconstruction
without expensive pretrained networks.
V. Will This Lead to Significant Benefits Soon?
(Short Answer: Yes, but in specific domains — not universally)
Areas likely to see near-term impact (1–3 years)
Architecture design for efficient vision systems
Dimensionality design (compress/expand) is easy to implement and can be rapidly adopted.Low-data and embedded AI systems
Training-free or lightly trained CNNs could be immediately useful in robotics, drones, and IoT perception.Computational neuroscience modeling
These results provide a major methodological shift, allowing scientists to model cortical properties using untrained networks.Neural decoding models
Faster, more biologically aligned models could help in medical and cognitive-neuroscience applications.
Areas where benefits will be slower (3–7+ years)
Replacing or reshaping foundation vision models
Despite architectural insights, Transformers dominate in scaling performance.
CNN-based cortical constraints will need to integrate with large-scale multi-modal training regimes.Influencing AGI architectures
Though useful, cortical alignment is not the only requirement for AGI; therefore, adoption will be partial.General safety and interpretability improvements
While promising, architectural improvements alone cannot solve emergent misalignment in large-scale systems.
Conclusion
These developments represent a major conceptual advance: the structure of convolutional architectures alone is sufficient to produce representations aligned with primate visual cortex, even without training. This challenges the longstanding belief that cortex-like representations arise from extensive learning on natural visual input, and it elevates the importance of architectural inductive biases as a primary explanatory factor for both biological and artificial vision systems.
The surprising and controversial findings — especially the spontaneous emergence of alignment and the critical role of spatial compression plus feature expansion — have important implications for AI design, computational neuroscience, robotics, and interpretability research. While these insights are unlikely to overturn the Transformer-dominated landscape in the immediate term, they will increasingly influence specialized AI applications, hybrid architectures, and neuroscientific methodologies.
Net assessment:
Significant benefits are likely in the near term for neuroscience, robotics, and efficient perception architectures, while transformative impact on mainstream foundation models may take longer to materialize.

Easy-to-Understand Essay: What It Means That Convolutional Networks Look Like the Brain “From the Start”
by ChatGPT-5.1
The research makes a simple but surprisingly powerful claim: some deep-learning models already look like the human and primate visual cortex even before they learn anything.
In other words, parts of the brain’s visual system seem to appear naturally inside certain artificial neural networks just because of how they are built.
This goes directly against the common belief in both neuroscience and AI that models become brain-like only after training on millions of images. The authors show that this assumption is incomplete — and in some cases, wrong.
This essay explains the findings in plain language, highlights the most surprising and controversial points, and outlines what these results may mean for future AI development.
1. What the Researchers Found (in simple terms)
The authors tested many different kinds of neural network architectures, all with random weights (no training, no learning). They compared how well these networks matched real brain activity in humans and monkeys viewing images.
The results were clear:
1. Convolutional neural networks (CNNs) naturally resemble the visual cortex
Even without training, CNNs showed patterns of activity that matched real visual cortex responses surprisingly well.
2. Two architectural features were essential
The brain-like representations only appeared when CNNs included:
Spatial compression
– This is done by pooling layers, which shrink spatial resolution as the network goes deeper.
– The visual system does something similar as signals move from the retina up into the brain.Feature expansion
– This means increasing the number of channels (filters) as the network gets deeper.
– Biological vision also expands the number of features as signals ascend the ventral stream.
These two steps — compressing space while expanding features — were the magic combination.
3. Other architectures don’t show this effect
Transformers, MLPs, and hybrid architectures did not end up looking like the visual cortex unless they were trained on huge datasets. Only CNNs showed alignment “de novo” — right out of the box.
This was true even when networks were extremely wide.
4. Small changes to the CNN architecture destroy the effect
When the researchers “lesioned” CNNs (for example, by disrupting weight sharing or removing pooling), the brain-like properties disappeared.
This suggests the effect isn’t accidental — it’s truly the architecture doing the work.
5. Training improves things, but the foundation is already there
Training largely refines a structure that is already pointing in the right direction.
2. Why This Is Surprising
Surprise #1: Training is not the main reason CNNs resemble the brain
For years, people assumed similarity came from learning the structure of the world.
This paper says: “Actually, their wiring already points in the right direction.”
Surprise #2: Random CNNs outperform trained models of other types
Transformers need massive training to look even slightly brain-like.
CNNs don’t.
Surprise #3: The brain’s visual system may also be largely shaped by architecture
This supports a new idea in neuroscience:
biology might not need tons of experience to organize the visual system — the structure of the cortex itself may do most of the work.
3. Why This Is Controversial
1. It challenges 10+ years of neuroscience-AI literature
A lot of influential work argued that brain-model similarity comes from task-driven training (e.g., “optimize for object recognition and you’ll look like IT cortex”).
These papers contradict that narrative.
2. It suggests some benchmarks overemphasize accidental similarities
If untrained CNNs already score well on brain-matching tasks, then training may not be the main reason models look biologically plausible.
3. It raises uncomfortable questions about Transformers
Transformers dominate modern AI, but biologically they may be much further from cortex-like processing — unless heavily trained.
4. It hints the brain may be more “engineered” than “learned”
This touches on long-running debates about nature vs. nurture in perception.
4. Valuable Insights for AI Developers
1. CNN architectural biases are still extremely useful
Even in the age of Transformers, CNNs offer:
natural locality
natural invariance
natural hierarchical structure
natural feature growth with depth
The paper reinforces that CNNs remain deeply elegant and efficient for vision.
2. You can build useful, brain-like models without massive training
This has practical implications for:
robotics
embedded devices
low-energy systems
security-sensitive applications
neuromorphic hardware
Training-free or minimally trained models could become more common.
3. Good architecture can replace some of the work of training
If you design the geometry of a network well, you may not need as much pretraining.
4. Better interpretability and safety
Models whose core representations come from architecture (rather than mysterious training dynamics) are:
more predictable
easier to analyze
easier to constrain
This is a major advantage for safety-critical AI.
5. How These Findings Could Shape the Future of AI
Near-term (1–3 years)
We can expect:
renewed interest in CNNs and CNN–Transformer hybrids
more low-data models in robotics
better tools for decoding brain activity
energy-efficient “architecture-first” vision systems
This is where the impact will be strongest.
Medium-term (3–7 years)
These insights may help redesign:
multimodal models
vision components inside LLMs
more biologically-grounded AGI approaches
But these changes will take time.
Long-term (7+ years)
This research could influence how we combine neuroscience and AI at a deeper level — maybe leading to models that learn and generalize more like brains.
6. Will This Lead to Big Benefits Soon?
Yes — but mainly for certain areas.
High near-term impact
brain modeling
robotics and embodied AI
efficient vision systems
interpretability research
neural decoding and medical imaging
Moderate or slow impact
replacing Transformers
large-scale multimodal foundation models
AGI research
safety architectures at the frontier scale
Transformers will not be replaced overnight.
But the design principles revealed here are likely to influence the next generation of hybrid systems.
Conclusion
The core message of the paper is simple: some neural network architectures naturally think a little like the visual cortex — even without training.
This reshapes how we understand both biological vision and machine vision. It reminds us that structure often matters as much as learning. And it gives us new tools to design AI systems that are more efficient, more interpretable and, in some domains, more human-like.
Over time, these ideas may help guide a shift toward architectures that blend the best of CNNs, Transformers and biology — but the biggest benefits will likely show up first in fields like neuroscience, robotics and low-power AI, where architectural bias matters more than scale.