SfN 2025 Nanosymposium

Deep neural networks have become the leading models for predicting neural responses in midātoāhigher visual areas of macaques and humans. However, nearāceiling prediction accuracy can mask whether these models truly capture the neuronal feature tuning or merely exploit spurious correlations. To address this, we leveraged feature accentuation, a technique allowing an encoding model to modify seed images to amplify the features it uses. By synthesizing images predicted to drive a neuron at specified levels, we tested whether encoding models could ācontrolā neural activity across its full dynamic rangeāan essential criterion for a true ādigital twin.ā
In the first session of our closedāloop paradigm, we recorded responses from five macaques to ā¼1,000 natural images in rapid serial visual presentation. Neurons in V3, V4, or anterior IT were recorded via floating microelectrode arrays or Neuropixel probes. Overnight, we fit encoding models by ridge regression on hiddenālayer activations from ten preātrained backbones (AlexNet, ResNet, and vision transformers (ViT), including adversarially trained versions). After evaluating accuracy on heldāout images, we selected each channelās bestāpredicting layer. For the most reliable channels, each model generated featureāaccentuated (FA) images at 11 target activation levelsāfrom below to beyond naturalāimage responsesāusing ten seed images.
In a second recording session, we presented these FA images back and measured actual neural responses, comparing them to model predictions. Although most models (except AlexNet) had similar prediction accuracy on heldāout natural images, their FA images differed substantially across backbones, reflecting diverse features used for prediction. Critically, adversarially trained ResNet and ViT consistently produced FA images that modulated real neurons in the intended parametric manner, with correlation for certain channels beyond 0.8. The FA images appeared more āshapeālike,ā whereas those from other models were more ātextureālike,ā suggesting that adversarial training promotes features better aligned with cortical neurons.
Overall, our results demonstrate that standard fitting and evaluation procedure can yield models with equivalent predictive scores yet rely on differentāand sometimes spuriousāfeatures. Closedāloop control is a more stringent test, revealing that adversarial training encourages shapeābased representations that better match true neural tuning. Our comparison paves the way for developing foundation models of visual cortex and predicting causal interventions for neural control.
San Diego, CA
San Diego, CA