AQAL: A Foundation Model Approach to Predicting Neurodiverse Brain Activity
Leeza Care Research & Development Foundation
Abstract
We present AQAL, a computational system that predicts how neurodiverse (autistic) brains process sensory stimuli by combining a proprietary multimodal brain encoding model (177M parameters) with a statistical neurodiverse transform derived from 871 resting-state fMRI scans from a large-scale multi-site autism neuroimaging consortium. AQAL maps text, audio, and video inputs onto 20,484 cortical surface vertices and generates both neurotypical and neurodiverse brain activation predictions in real time. Our connectivity analysis identifies 820 statistically significant inter-regional differences (p < 0.05) between ASD and typically-developing groups across 100 cortical parcels, with the limbic temporal pole, default mode network, and visual cortex showing the largest effect sizes. We further introduce a sensory profiling module that quantifies divergence across seven canonical brain networks, enabling practical applications in accessibility auditing and personalized accommodation design.
Contents
1.Introduction
Autism spectrum conditions affect approximately 1 in 36 children worldwide (CDC, 2023). Understanding how autistic brains process sensory information differently from neurotypical brains is critical for designing inclusive environments, educational strategies, and clinical interventions. However, functional neuroimaging studies are expensive, slow, and inaccessible to most practitioners and families.
Recent advances in brain encoding models — neural networks trained to predict brain activity from sensory stimuli — offer a new paradigm. State-of-the-art multimodal transformers can now predict cortical surface activity from vision, audition, and language inputs with meaningful accuracy across subjects.
We asked: Can a brain encoding model, combined with population-level connectivity data from autism neuroimaging, predict how a neurodiverse brain would respond to arbitrary stimuli?
AQAL is our answer. Rather than requiring individual brain scans, AQAL uses a proprietary statistical transform trained on a large multi-site fMRI corpus to convert neurotypical brain predictions into neurodiverse variants, making neuroscience-informed predictions accessible through a web API.
1.1 Contributions
- A CPU-based statistical transform that converts neurotypical predictions to neurodiverse predictions using connectivity effect sizes from 871 subjects across 20 clinical sites.
- A connectivity analysis pipeline identifying 820 significant inter-regional differences between ASD and TD groups across 4,950 tested connections.
- A sensory profiling system that maps brain-level divergence onto seven functional networks (visual, auditory, motor, language, social, default mode, salience).
- A publicly accessible platform (API + web interface) enabling real-time neurodiverse brain prediction from text, audio, or video input.
3.Methods
3.1 System Architecture
Stimulus (text / audio / video)
│
▼
┌───────────────────────┐
│ Feature Extraction │ Proprietary multimodal encoders
└──────────┬────────────┘
│
▼
┌───────────────────────┐
│ AQAL Brain Encoder │ 8-layer transformer, 177M params
│ (NT Prediction) │ → 20,484 vertices × T timesteps
└──────────┬────────────┘
│
▼
┌───────────────────────┐
│ ND Transform │ Trained on 871 multi-site subjects
│ (ASD Prediction) │ Per-vertex scale + shift
└──────────┬────────────┘
│
▼
┌───────────────────────┐
│ Analysis & │ Divergence metrics, sensory profile,
│ Visualization │ connectivity mapping, interpretation
└───────────────────────┘3.2 AQAL Brain Encoder
The AQAL brain encoder is a proprietary multimodal transformer trained to predict cortical surface activity from sensory inputs. The model architecture consists of:
- Feature extractors: Dedicated encoders for text (6 layers), video (2 layers), and audio (2 layers) built on state-of-the-art foundation models
- Feature projectors: Per-modality MLPs mapping to a common hidden dimension of 1,152
- Temporal smoothing: Gaussian kernel (size 9) for hemodynamic response alignment
- Transformer encoder: 8 layers, 18 heads, head dimension 64, FF dimension 3,072
- Subject layers: Low-rank output adaptation (bottleneck 2,048) with 0.1 dropout
- Output space: 20,484 vertices on a standard cortical surface mesh (10,242 per hemisphere)
The model operates at 2 Hz temporal resolution. Modality dropout of 0.3 enables predictions from any subset of input modalities.
3.3 Training Dataset
We trained the neurodiverse transform on a large-scale multi-site autism neuroimaging corpus comprising 871 resting-state fMRI scans:
| Group | N | Sites |
|---|---|---|
| ASD | 403 | 20 |
| TD | 468 | 20 |
All scans were preprocessed using a standard fMRI preprocessing pipeline with band-pass filtering (0.01–0.1 Hz) and no global signal regression. We extracted mean timeseries from 100 cortical parcels using a validated cortical parcellation, computed Pearson correlation matrices, and applied Fisher's z-transform for variance stabilization, producing 4,950 unique connectivity features per subject.
3.4 Statistical Group Comparison
For each of the 4,950 unique connectivity pairs, we performed an independent-samples t-test comparing ASD and TD groups. 820 of 4,950 connections (16.6%) were significant at p < 0.05.
The most affected regions ranked by normalized effect size:
| Rank | Region | Network | Effect |
|---|---|---|---|
| 1 | Limbic Temporal Pole R | Limbic | 1.000 |
| 2 | Limbic Temporal Pole L | Limbic | 0.857 |
| 3 | Visual 9 L | Visual | 0.551 |
| 4 | Default Mode Temporal 2 R | Default Mode | 0.549 |
| 5 | Default Mode Temporal 2 L | Default Mode | 0.527 |
These findings are consistent with the autism neuroimaging literature, which identifies limbic hyperconnectivity and default mode network alterations as hallmarks of ASD (Hull et al., 2017).
3.5 Neurodiverse Transform
We derive a proprietary per-vertex transform that converts neurotypical predictions to neurodiverse predictions. For each of 20,484 cortical vertices v:
- Identify the cortical ROI r(v) to which vertex v belongs.
- Compute the normalized effect size er for that ROI from connectivity t-tests.
- Compute scale and shift factors using a proprietary mapping function.
- Apply: ŷvND = ŷvNT × scalev + shiftv
Of 20,484 vertices, 18,067 (88.3%) are affected by the transform, reflecting the widespread but heterogeneous nature of connectivity differences in autism.
3.6 Sensory Profiling
We map vertex-level divergence between NT and ND predictions onto seven canonical functional brain networks:
| Network | Function | Brain Areas |
|---|---|---|
| Visual | Processing sight | Primary and secondary visual cortex |
| Auditory | Processing sound | Auditory cortex, belt regions |
| Language | Speech and words | Broca's and Wernicke's areas |
| Motor | Movement and body | Primary motor and somatosensory cortex |
| Social | Understanding others | Superior temporal regions |
| Default Mode | Self-reflection | Posterior cingulate, medial prefrontal |
| Salience | Filtering what matters | Insula, anterior cingulate |
Scores are normalized to [0, 1] by dividing by the maximum network divergence, providing an interpretable summary of which processing domains are most differently activated.
3.7 Infrastructure
| Component | Technology |
|---|---|
| Brain model | AQAL Encoder (proprietary, PyTorch) |
| API server | Cloud VM (8 cores, 32 GB RAM) |
| Frontend | Next.js on Vercel |
| Interpretation | LLM-powered (GPT) |
All inference runs on CPU. No GPU is required.
4.Results
4.1 Connectivity Analysis
Analysis of 871 subjects across 20 clinical sites revealed 820 significant connections out of 4,950 tested (16.6%, p < 0.05). The limbic system showed the largest effect sizes, consistent with known emotional regulation differences. Default mode network showed bilateral temporal alterations. Visual cortex alterations align with documented sensory processing variations.
4.2 Network-Level Differences
| Network | Difference | Direction |
|---|---|---|
| Limbic | Highest | Hyperconnectivity |
| Default Mode | High | Altered connectivity |
| Visual | Moderate–High | Modified processing |
| Salience | Moderate | Altered filtering |
| Dorsal Attention | Moderate | Shifted focus |
| Control | Low–Moderate | Mild executive differences |
| Somatomotor | Low–Moderate | Subtle motor differences |
4.3 Transform Coverage
The neurodiverse transform affects 18,067 of 20,484 cortical vertices (88.3%), with scale factors ranging from 0.85 to 1.15 and shift factors from \u22120.05 to 0.01. The transform is denser in temporal and limbic regions and sparser in primary motor cortex, matching the known neuroanatomy of autism.
4.4 Real-Time Performance
| Operation | Latency |
|---|---|
| Text → events (TTS + ASR) | ~3–5s |
| AQAL encoder inference (10 timesteps) | ~8–15s |
| ND transform application | < 100ms |
| Surface rendering (4 views × 10 steps) | ~5–8s |
| Total end-to-end | ~20–30s |
5.Applications
5.1 Sensory Accessibility Auditing
AQAL can process video of physical spaces (classrooms, offices, retail environments) and generate second-by-second sensory stress predictions for neurodiverse individuals, including flagged high-stress moments and actionable accessibility recommendations.
5.2 Educational Accommodation Design
By profiling how specific educational content differently activates neurodiverse brain networks, AQAL can inform content pacing, optimal break timing, and modality-specific accommodations.
5.3 Clinical Communication
The sensory profile output provides a quantitative, brain-data-grounded framework for communicating how a neurodiverse individual might experience specific stimuli — replacing subjective descriptions with measurable network-level predictions.
6.Limitations
- No GPU fine-tuning. The neurodiverse transform is a statistical approximation, not a directly fine-tuned neural network. GPU fine-tuning on neurodiverse fMRI data would likely improve accuracy but was not feasible due to compute constraints.
- Indirect mapping. The transform maps connectivity-level group differences onto vertex-level predictions. This assumes resting-state connectivity differences generalize to task-evoked activity — a reasonable but imperfect assumption.
- Group-level, not individual. AQAL produces an average neurodiverse prediction. Autism is a spectrum, and individual variation is substantial.
- Uncorrected p-values. The 820 significant connections are reported at p < 0.05 without multiple comparison correction. With FDR correction, the number decreases.
- Dataset limitations. The training corpus is heterogeneous (20 sites, varying protocols) and skewed toward males. Results may not generalize equally to all populations.
7.Future Work
- GPU fine-tuning of the AQAL encoder directly on neurodiverse fMRI data
- Individual calibration via a 5-minute sensory assessment to personalize predictions
- Expanded training data from additional consortia to reach 10,000+ subjects
- Additional neurodivergent conditions (ADHD, sensory processing disorder, anxiety)
- EEG integration for real-time, portable brain monitoring
- Clinical validation studies comparing AQAL predictions against observed behavioral responses
8.Conclusion
AQAL demonstrates that combining a state-of-the-art multimodal brain encoding model with population-level autism neuroimaging data can produce meaningful neurodiverse brain predictions from arbitrary sensory stimuli. While the current statistical transform is an approximation, it makes neuroscience-informed predictions accessible without requiring individual brain scans or GPU infrastructure. The identification of 820 significant connectivity differences across 871 subjects provides a robust empirical foundation, and the system's real-time performance enables practical applications in accessibility, education, and clinical communication.
By releasing the platform publicly, we aim to make computational neurodiversity research accessible to researchers, educators, clinicians, and families who lack access to neuroimaging facilities.
References
CDC. (2023). Autism Spectrum Disorder: Data & Statistics. Centers for Disease Control and Prevention.
de Heer, W. A., Huth, A. G., Griffiths, T. L., Gallant, J. L., & Theunissen, F. E. (2017). The hierarchical cortical organization of human speech processing. Journal of Neuroscience, 37(27), 6539–6557.
Holiga, S., et al. (2019). Patients with autism spectrum disorders display reproducible functional connectivity alterations. Science Translational Medicine, 11(481).
Hull, J. V., et al. (2017). Resting-state functional connectivity in autism spectrum disorders: A review. Frontiers in Psychiatry, 7, 205.
Huth, A. G., de Heer, W. A., Griffiths, T. L., Theunissen, F. E., & Gallant, J. L. (2016). Natural speech reveals the semantic maps that tile human cerebral cortex. Nature, 532(7600), 453–458.
Ortega Caro, J., et al. (2023). BrainLM: A Foundation Model for Brain Activity Recordings. arXiv preprint arXiv:2311.00656.
Thomas, A. W., et al. (2023). Self-supervised learning of brain dynamics from broad neuroimaging data. Advances in Neural Information Processing Systems, 36.
Appendix A: Model Parameters
| Parameter | Value |
|---|---|
| Total parameters | 177M |
| Hidden dimension | 1,152 |
| Transformer layers | 8 |
| Attention heads | 18 |
| Head dimension | 64 |
| Feedforward dimension | 3,072 |
| Output vertices | 20,484 |
| Temporal resolution | 2 Hz |
| Low-rank bottleneck | 2,048 |
| Modality dropout | 0.3 |
| Subject dropout | 0.1 |
Appendix B: Training Corpus Demographics
| Metric | ASD | TD |
|---|---|---|
| N | 403 | 468 |
| Mean age | 16.9 ± 7.8 | 16.6 ± 7.0 |
| Male (%) | 86.6 | 84.2 |
| Sites | 20 | 20 |
Appendix C: System Availability
| Resource | URL |
|---|---|
| Landing page | mind.new |
| NeuroBrain | neuro.mind.new |
| Sensory Audit | sensory.mind.new |