pith. machine review for the scientific record. sign in

arxiv: 2605.12485 · v1 · submitted 2026-05-12 · 🧬 q-bio.NC · q-bio.QM

Recognition: no theorem link

Letting the neural code speak: Automated characterization of monkey visual neurons through human language

Andreas S. Tolias, Katrin Franke, Nikos Karantzas, Sophia Sanborn, Surya Ganguli, Tamar Rott Shaham, Vedang Lad

Pith reviewed 2026-05-13 02:05 UTC · model grok-4.3

classification 🧬 q-bio.NC q-bio.QM
keywords visual cortexmacaque neuronsneural selectivitylanguage descriptionsdigital twinssemantic hypothesesV1V4
0
0 comments X

The pith

Natural language descriptions capture the selectivity of most neurons in macaque V1 and V4.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper establishes that concise semantic descriptions in human language can characterize what activates or suppresses individual neurons in monkey visual cortex, replacing the need for custom mathematical models in higher areas. It uses digital twins of V1 and V4 to turn strong and weak response images into text captions, form hypotheses about features like edges or color-texture conjunctions, generate new images, and check those images against the models. This process succeeds for nearly all tested V4 neurons and many in V1, with language hypotheses producing extreme responses far beyond random images. A sympathetic reader would care because it offers a scalable, human-interpretable way to understand neural codes where traditional approaches fall short.

Core claim

Across macaque V1 and V4, the selectivity of most neurons is captured by concise, verifiable semantic descriptions. Using digital twins, the method translates high- and low-activating images into dense captions, generates a semantic hypothesis and synthesized images, and verifies the hypothesis in silico. In V4, images from activating and suppressing hypotheses drove 96.1% of neurons above the 95th and 97.6% below the 5th percentile of natural-image responses, respectively.

What carries the argument

The closed-loop framework that converts neuron responses into language hypotheses via digital twins, then renders those hypotheses back into images for in-silico verification.

If this is right

  • V4 neurons respond to conjunctions of form, color, and texture that language can name, while V1 responses align more with oriented edges and spatial frequency.
  • Representational similarity analysis shows vision embeddings align more closely with neural activity than language embeddings, yet rendering hypotheses back to images recovers much of the lost alignment.
  • The method produces testable predictions at scale without requiring new biological experiments for initial hypothesis generation.
  • Linguistic compression of neural selectivity is lossy but remains semantically faithful for verification purposes.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same language-based loop could be applied to areas beyond V4 where no simple mathematical models exist, potentially revealing higher-order feature combinations.
  • If the digital twins generalize well, this framework might accelerate discovery by letting researchers query neural populations with natural language rather than exhaustive image searches.
  • The partial mismatch between language and vision embeddings suggests that some visual features driving neurons may resist concise verbal description and require additional modalities for full capture.

Load-bearing premise

The digital-twin models of V1 and V4 accurately reproduce how real biological neurons respond to the novel synthetic images generated from the language hypotheses.

What would settle it

Presenting the language-generated activating and suppressing images to real V4 neurons and finding that they fail to drive responses above the 95th or below the 5th percentile of natural images would falsify the claim that semantic descriptions capture selectivity.

Figures

Figures reproduced from arXiv: 2605.12485 by Andreas S. Tolias, Katrin Franke, Nikos Karantzas, Sophia Sanborn, Surya Ganguli, Tamar Rott Shaham, Vedang Lad.

Figure 1
Figure 1. Figure 1: Framework for translating neural selectivity into interpretable semantic hypotheses. The pipeline consists of three stages: Translate: Each image is converted into a detailed textual description using Gemini 3.0 Pro. To evaluate the fidelity of this image-to-text translation, we regenerate images from the captions using a text-to-image model and quantify correspondence to the original image in an image-sim… view at source ↗
Figure 2
Figure 2. Figure 2: Translation and faithfulness of image-to-text descriptions. The Translate stage of our framework converts input images into detailed captions via Gemini 3.0 Pro and assesses faithfulness by comparing caption-conditioned reconstructions to the originals in DINOv3 embedding space. (a) Translate: Area V4. Given an input image (top left), Gemini 3.0 Pro generates a detailed, multi-sentence caption describing t… view at source ↗
Figure 3
Figure 3. Figure 3: Deriving semantic hypotheses from neurons in macaque visual cortex. For each V1 and V4 neuron, extreme-response images are identified from a large naturalistic image dataset via a functional digital twin. For neurons with baseline activity, we extract both top- and bottom-activating images and distill each set separately into an excitatory and a suppressive semantic hypothesis; for sparse neurons, we extra… view at source ↗
Figure 4
Figure 4. Figure 4: Area V4: Closed-loop verification of semantic hypotheses using generative stimuli and spatial opti￾mization. Top: A generated semantic hypothesis for an example V4 neuron is expanded into multiple diverse text prompts, which are then rendered into novel images using a text-to-image model. These generated images resemble the neuron’s most-activating natural images, capturing core feature conjunctions such a… view at source ↗
Figure 5
Figure 5. Figure 5: Area V1: Closed-loop verification of semantic hypotheses using generative stimuli and spatial opti￾mization. Semantic hypotheses successfully generate stimuli that drive neurons above the random baseline, confirming that the pipeline generalizes across the visual hierarchy. The smaller gain from spatial optimization relative to V4 quantifies the expected gradient: language is a coarser coordinate system fo… view at source ↗
Figure 7
Figure 7. Figure 7: Semantic structure of neural selectivity revealed through population activity clustering. Left: UMAP embedding of V4 neurons clustered by population activity similarity, annotated with nouns and adjectives extracted from the first sentence of each neuron’s semantic hypothesis. Large-scale neighborhoods exhibit smooth transitions in both visual content and descriptive language, from eyes and circular organi… view at source ↗
read the original abstract

Understanding what individual neurons encode is a core question in neuroscience. In primary visual cortex (V1), mathematical models (e.g., Gabor functions) capture neural selectivity, but no comparable framework exists for higher areas. We show that natural language can fill this role: across macaque V1 and V4, the selectivity of most neurons is captured by concise, verifiable semantic descriptions. Using digital twins of V1 and V4, we develop a closed-loop framework that translates each neuron's high- and low-activating images into dense captions, generates a semantic hypothesis and synthesized images, and verifies the hypothesis in silico. Descriptions range from oriented edges and spatial frequency in V1 to conjunctions of form, color, and texture in V4. In V4, images generated from activating and suppressing hypotheses drove 96.1% of neurons above the 95th and 97.6% below the 5th percentile of natural-image responses, respectively (vs. ~10\% for random images); V1 activation results matched V4, while V1 suppression was less describable in language. Representational similarity analysis reveals partial alignment between neural activity, vision embeddings, and language embeddings, with vision most aligned to neural activity; alignment lost in the text bottleneck is recovered when hypotheses are rendered back into images, showing that linguistic compression is lossy yet semantically faithful. Together, these results show that combining generative models with neural digital twins enables interpretable, testable descriptions of neural function at scale, toward agentic scientific discovery.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript presents a closed-loop framework using digital twins of macaque V1 and V4 neurons to automatically generate concise natural-language semantic descriptions of each neuron's selectivity. High- and low-activating natural images are captioned, a semantic hypothesis is formed via language models, new images are synthesized from the hypothesis, and the hypothesis is verified in silico by querying the digital twins; the authors report that activating images drive 96.1% of V4 neurons above the 95th percentile of natural-image responses and suppressing images drive 97.6% below the 5th percentile (versus ~10% for random images), with comparable activation but weaker suppression results in V1. Representational similarity analysis is used to compare neural activity, vision embeddings, and language embeddings.

Significance. If the digital twins prove reliable on the synthetic images, the work would offer a scalable route to interpretable characterizations of visual selectivity in higher areas where Gabor-style models are inadequate, and would illustrate how generative language and image models can be combined with neural digital twins for automated hypothesis generation and testing. The partial alignment results between modalities, with vision closest to neural activity and recovery upon re-rendering, provide additional insight into semantic compression.

major comments (3)
  1. [Results section describing V4 activation and suppression verification] The central performance claims (96.1% activation and 97.6% suppression in V4) rest entirely on in-silico queries of the digital twins applied to LLM-generated synthetic images that lie outside the natural-image distribution on which the twins were trained. No section reports direct biological recordings on these novel images, nor any quantitative metric (e.g., held-out correlation, response-distribution match, or generalization error) confirming that twin predictions remain faithful for the particular conjunctions of form, color, and texture produced by the language model.
  2. [Methods section on digital-twin construction] The manuscript provides no quantitative details on the digital-twin models themselves: number of neurons recorded and modeled, training data composition, architecture, regularization, or performance on held-out natural images. Without these, the reliability of the in-silico verification step cannot be assessed, and the reported percentages may reflect model idiosyncrasies rather than biological selectivity.
  3. [Results section on representational similarity analysis] The representational similarity analysis claims that 'alignment lost in the text bottleneck is recovered when hypotheses are rendered back into images' and that this demonstrates semantic faithfulness. The specific distance metrics, number of stimuli, statistical controls, and comparison to null models are not detailed enough to evaluate whether the recovery is attributable to semantic content rather than low-level image statistics.
minor comments (2)
  1. [Abstract] The abstract states 'vs. ~10% for random images' without specifying the exact percentile thresholds, number of random images, or statistical test; this comparison should be made explicit.
  2. [Figure captions] Figure legends and captions should explicitly state the number of neurons, number of images per condition, and exact percentile definitions used for the activation and suppression results.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive and detailed comments. We address each major point below, providing clarifications and committing to revisions that strengthen the manuscript without altering its core claims.

read point-by-point responses
  1. Referee: [Results section describing V4 activation and suppression verification] The central performance claims (96.1% activation and 97.6% suppression in V4) rest entirely on in-silico queries of the digital twins applied to LLM-generated synthetic images that lie outside the natural-image distribution on which the twins were trained. No section reports direct biological recordings on these novel images, nor any quantitative metric (e.g., held-out correlation, response-distribution match, or generalization error) confirming that twin predictions remain faithful for the particular conjunctions of form, color, and texture produced by the language model.

    Authors: We acknowledge that the verification relies on in-silico queries and that direct biological recordings on the LLM-generated synthetic images are not reported. The framework is intentionally designed for scalable automated testing via digital twins rather than exhaustive new recordings for each hypothesis. We will add quantitative validation metrics for the twins (held-out correlation and response-distribution statistics on natural images) and explicitly discuss the generalization assumption as a limitation, with suggestions for future wet-lab confirmation. revision: partial

  2. Referee: [Methods section on digital-twin construction] The manuscript provides no quantitative details on the digital-twin models themselves: number of neurons recorded and modeled, training data composition, architecture, regularization, or performance on held-out natural images. Without these, the reliability of the in-silico verification step cannot be assessed, and the reported percentages may reflect model idiosyncrasies rather than biological selectivity.

    Authors: We agree that these details are essential for assessing reliability. In the revised Methods section we will report the number of V1 and V4 neurons recorded and modeled, the size and composition of the natural-image training sets, the model architectures and regularization procedures, and performance metrics (e.g., held-out Pearson correlation and response-distribution match) on natural images. revision: yes

  3. Referee: [Results section on representational similarity analysis] The representational similarity analysis claims that 'alignment lost in the text bottleneck is recovered when hypotheses are rendered back into images' and that this demonstrates semantic faithfulness. The specific distance metrics, number of stimuli, statistical controls, and comparison to null models are not detailed enough to evaluate whether the recovery is attributable to semantic content rather than low-level image statistics.

    Authors: We will expand both the Methods and Results sections to specify the distance metrics (e.g., cosine similarity on normalized embeddings), the exact number of stimuli per comparison, the statistical controls employed, and the null-model procedures (including shuffled and low-level-statistic-matched controls). These additions will allow readers to confirm that the reported recovery reflects semantic rather than low-level image properties. revision: yes

Circularity Check

1 steps flagged

In-silico verification of semantic hypotheses is performed by querying digital twins fitted to the same neural data

specific steps
  1. fitted input called prediction [Abstract]
    "Using digital twins of V1 and V4, we develop a closed-loop framework that translates each neuron's high- and low-activating images into dense captions, generates a semantic hypothesis and synthesized images, and verifies the hypothesis in silico. ... In V4, images generated from activating and suppressing hypotheses drove 96.1% of neurons above the 95th and 97.6% below the 5th percentile of natural-image responses, respectively (vs. ~10% for random images)"

    The percentile-driving claims are computed by evaluating the synthesized images inside the digital twins. Because the twins are fitted to the same neural data used to identify the original high/low-activating images and to generate the hypotheses, the high success rates quantify agreement with the fitted model rather than an external measurement of biological selectivity.

full rationale

The paper's central quantitative result (96.1% and 97.6% of V4 neurons driven above/below response percentiles) is obtained by feeding LLM-generated synthetic images into digital-twin models whose parameters were fit to the original neural recordings. This makes the reported verification a measure of consistency inside the fitted model rather than an independent biological test, matching the fitted-input-called-prediction pattern. No other circular steps (self-citations, self-definitional equations, or imported uniqueness theorems) appear in the provided text.

Axiom & Free-Parameter Ledger

2 free parameters · 2 axioms · 0 invented entities

The central claim rests on the existence and fidelity of digital-twin models trained on macaque V1/V4 data, the ability of vision-language models to produce faithful captions and image generations, and the assumption that language can serve as a lossless-enough compression for neural selectivity.

free parameters (2)
  • Digital-twin model parameters
    Parameters of the V1 and V4 digital twins are fitted to neural responses; their exact count and training procedure are not stated in the abstract.
  • Captioning and generation model hyperparameters
    Choices inside the vision-language models used for captioning and image synthesis are not enumerated.
axioms (2)
  • domain assumption Digital twins accurately predict responses to novel synthetic images outside the training distribution
    Invoked when the paper treats in-silico verification as evidence for biological selectivity.
  • domain assumption Natural language is sufficiently expressive to capture the selectivity of V1 and V4 neurons
    Stated as the core premise that language can fill the role previously played by mathematical models.

pith-pipeline@v0.9.0 · 5606 in / 1684 out tokens · 41302 ms · 2026-05-13T02:05:44.662959+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

112 extracted references · 112 canonical work pages · 2 internal anchors

  1. [1]

    2026 , eprint=

    Revisiting the Platonic Representation Hypothesis: An Aristotelian View , author=. 2026 , eprint=

  2. [2]

    Cell , volume=

    Evolving images for visual neurons using a deep generative network reveals coding principles and neuronal preferences , author=. Cell , volume=. 2019 , publisher=

  3. [3]

    Interpreting the retinal neural code for natural scenes: From computations to neurons

    Maheswaranathan, Niru and McIntosh, Lane T and Tanaka, Hidenori and Grant, Satchel and Kastner, David B and Melander, Joshua B and Nayebi, Aran and Brezovec, Luke E and Wang, Julia H and Ganguli, Surya and Baccus, Stephen A. Interpreting the retinal neural code for natural scenes: From computations to neurons. Neuron

  4. [4]

    Neural representational geometry underlies few-shot concept learning

    Sorscher, Ben and Ganguli, Surya and Sompolinsky, Haim. Neural representational geometry underlies few-shot concept learning. Proc. Natl. Acad. Sci. U. S. A

  5. [5]

    Distill , volume=

    Multimodal neurons in artificial neural networks , author=. Distill , volume=

  6. [6]

    Nature Neuroscience , volume=

    Semantic reconstruction of continuous language from non-invasive brain recordings , author=. Nature Neuroscience , volume=. 2023 , publisher=

  7. [7]

    Proceedings of the National Academy of Sciences , volume=

    The neural architecture of language: Integrative modeling converges on predictive processing , author=. Proceedings of the National Academy of Sciences , volume=

  8. [8]

    arXiv preprint arXiv:2510.02182 , year=

    Uncovering semantic selectivity of latent groups in higher visual cortex with mutual information-guided diffusion , author=. arXiv preprint arXiv:2510.02182 , year=

  9. [9]

    Frontiers in Handwriting Recognition (ICFHR), 2014 14th International Conference on , pages=

    Real-time segmentation of on-line handwritten arabic script , author=. Frontiers in Handwriting Recognition (ICFHR), 2014 14th International Conference on , pages=. 2014 , organization=

  10. [10]

    Jones, J. P. and Palmer, L. A. , journal=. An evaluation of the two-dimensional

  11. [11]

    Journal of the Optical Society of America A , volume=

    Uncertainty relation for resolution in space, spatial frequency, and orientation optimized by two-dimensional visual cortical filters , author=. Journal of the Optical Society of America A , volume=

  12. [12]

    Nature , volume=

    Emergence of simple-cell receptive field properties by learning a sparse code for natural images , author=. Nature , volume=. 1996 , publisher=

  13. [13]

    Annual Review of Neuroscience , volume=

    Natural image statistics and neural representation , author=. Annual Review of Neuroscience , volume=. 2001 , publisher=

  14. [14]

    Vision Research , volume=

    Quantitative analysis of cat retinal ganglion cell response to visual stimuli , author=. Vision Research , volume=. 1965 , publisher=

  15. [15]

    Journal of Neurophysiology , volume=

    Coding visual images of objects in the inferotemporal cortex of the macaque monkey , author=. Journal of Neurophysiology , volume=

  16. [16]

    and Connor, C

    Pasupathy, A. and Connor, C. E. , journal=. Shape representation in area

  17. [17]

    and Connor, C

    Pasupathy, A. and Connor, C. E. , journal=. Population coding of shape in area. 2002 , publisher=

  18. [18]

    Science , volume=

    A cortical region consisting entirely of face-selective cells , author=. Science , volume=. 2006 , publisher=

  19. [19]

    Annual Review of Neuroscience , volume=

    Mechanisms of face perception , author=. Annual Review of Neuroscience , volume=. 2010 , publisher=

  20. [20]

    Nature Neuroscience , volume=

    Metamers of the ventral stream , author=. Nature Neuroscience , volume=. 2011 , publisher=

  21. [21]

    Nature Neuroscience , volume=

    Underlying principles of visual shape selectivity in posterior inferotemporal cortex , author=. Nature Neuroscience , volume=. 2004 , publisher=

  22. [22]

    Cold Spring Harbor Symposia on Quantitative Biology , volume=

    Representation of naturalistic image structure in the primate visual cortex , author=. Cold Spring Harbor Symposia on Quantitative Biology , volume=. 2014 , publisher=

  23. [23]

    , journal=

    Oliver, Michael and Winter, Michele and Dupré la Tour, Tom and Eickenberg, Michael and Gallant, Jack L. , journal=. A biologically-inspired hierarchical convolutional energy model predicts. 2024 , doi=

  24. [24]

    Nature Machine Intelligence , volume=

    Better models of human high-level visual cortex emerge from natural language supervision with a large and diverse dataset , author=. Nature Machine Intelligence , volume=. 2023 , publisher=

  25. [25]

    and Wehbe, Leila , booktitle=

    Luo, Andrew and Henderson, Margot and Tarr, Michael J. and Wehbe, Leila , booktitle=

  26. [26]

    2025 , month=

    Wasserman, Navve and Cosarinsky, Matias and Golbari, Yuval and Oliva, Aude and Torralba, Antonio and Rott Shaham, Tamar and Irani, Michal , journal=. 2025 , month=

  27. [27]

    International Conference on Learning Representations (ICLR) , year=

    Rethinking Language-Alignment in Human Visual Cortex with Syntax Manipulation and Word Models , author=. International Conference on Learning Representations (ICLR) , year=

  28. [28]

    Nature Machine Intelligence , volume=

    High-level visual representations in the human brain are aligned with large language models , author=. Nature Machine Intelligence , volume=. 2025 , publisher=

  29. [29]

    Nature , volume=

    Invariant visual representation by single neurons in the human brain , author=. Nature , volume=. 2005 , publisher=

  30. [30]

    Proceedings of the National Academy of Sciences , volume=

    Human single-neuron responses at the threshold of conscious recognition , author=. Proceedings of the National Academy of Sciences , volume=. 2008 , publisher=

  31. [31]

    Frontiers in Systems Neuroscience , volume=

    Representational similarity analysis -- connecting the branches of systems neuroscience , author=. Frontiers in Systems Neuroscience , volume=. 2008 , publisher=

  32. [32]

    Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition , pages=

    Interpretable convolutional neural networks , author=. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition , pages=

  33. [33]

    Proceedings of the National Academy of Sciences , volume=

    Understanding the role of individual units in a deep neural network , author=. Proceedings of the National Academy of Sciences , volume=. 2020 , publisher=

  34. [34]

    Proceedings of the European Conference on Computer Vision , pages=

    Interpretable basis decomposition for visual explanation , author=. Proceedings of the European Conference on Computer Vision , pages=. 2018 , organization=

  35. [35]

    Proceedings of the National Academy of Sciences , volume=

    Performance-optimized hierarchical models predict neural responses in higher visual cortex , author=. Proceedings of the National Academy of Sciences , volume=. 2014 , publisher=

  36. [36]

    The UK Biobank resource with deep phenotyping and genomic data

    Bycroft, Clare and Freeman, Colin and Petkova, Desislava and Band, Gavin and Elliott, Lloyd T and Sharp, Kevin and Motyer, Allan and Vukcevic, Damjan and Delaneau, Olivier and O'Connell, Jared and Cortes, Adrian and Welsh, Samantha and Young, Alan and Effingham, Mark and McVean, Gil and Leslie, Stephen and Allen, Naomi and Donnelly, Peter and Marchini, Jo...

  37. [37]

    Highly accurate protein structure prediction with AlphaFold

    Jumper, John and Evans, Richard and Pritzel, Alexander and Green, Tim and Figurnov, Michael and Ronneberger, Olaf and Tunyasuvunakool, Kathryn and Bates, Russ and Žídek, Augustin and Potapenko, Anna and Bridgland, Alex and Meyer, Clemens and Kohl, Simon A A and Ballard, Andrew J and Cowie, Andrew and Romera-Paredes, Bernardino and Nikolov, Stanislav and J...

  38. [38]

    Foundation model of neural activity predicts response to new stimulus types

    Wang, Eric Y and Fahey, Paul G and Ding, Zhuokun and Papadopoulos, Stelios and Ponder, Kayla and Weis, Marissa A and Chang, Andersen and Muhammad, Taliah and Patel, Saumil and Ding, Zhiwei and Tran, Dat and Fu, Jiakun and Schneider-Mizell, Casey M and MICrONS Consortium and Reid, R Clay and Collman, Forrest and da Costa, Nuno Maçarico and Franke, Katrin a...

  39. [39]

    Accurate medium-range global weather forecasting with 3D neural networks

    Bi, Kaifeng and Xie, Lingxi and Zhang, Hengheng and Chen, Xin and Gu, Xiaotao and Tian, Qi. Accurate medium-range global weather forecasting with 3D neural networks. Nature

  40. [40]

    High-speed, cortex-wide volumetric recording of neuroactivity at cellular resolution using light beads microscopy

    Demas, Jeffrey and Manley, Jason and Tejera, Frank and Barber, Kevin and Kim, Hyewon and Traub, Francisca Martínez and Chen, Brandon and Vaziri, Alipasha. High-speed, cortex-wide volumetric recording of neuroactivity at cellular resolution using light beads microscopy. Nat. Methods

  41. [41]

    Neuropixels 2.0: A miniaturized high-density probe for stable, long-term brain recordings

    Steinmetz, Nicholas A and Aydin, Cagatay and Lebedeva, Anna and Okun, Michael and Pachitariu, Marius and Bauza, Marius and Beau, Maxime and Bhagat, Jai and Böhm, Claudia and Broux, Martijn and Chen, Susu and Colonell, Jennifer and Gardner, Richard J and Karsh, Bill and Kloosterman, Fabian and Kostadinov, Dimitar and Mora-Lopez, Carolina and O'Callaghan, J...

  42. [42]

    2024 , eprint=

    The Platonic Representation Hypothesis , author=. 2024 , eprint=

  43. [43]

    2024 , eprint=

    The Linear Representation Hypothesis and the Geometry of Large Language Models , author=. 2024 , eprint=

  44. [44]

    2022 , eprint=

    Toy Models of Superposition , author=. 2022 , eprint=

  45. [45]

    Functional connectomics spanning multiple areas of mouse visual cortex

    MICrONS Consortium. Functional connectomics spanning multiple areas of mouse visual cortex. Nature

  46. [46]

    and Hong, Ha and Yamins, Daniel L

    Cadieu, Charles F. and Hong, Ha and Yamins, Daniel L. K. and Pinto, Nicolas and Ardila, Diego and Solomon, Ethan A. and Majaj, Najib J. and DiCarlo, James J. , journal=. Deep neural networks rival the representation of primate. 2014 , publisher=

  47. [47]

    Science , volume=

    Neural population control via deep image synthesis , author=. Science , volume=. 2019 , publisher=

  48. [48]

    Nature Neuroscience , volume=

    Inception loops discover what excites neurons most using deep predictive models , author=. Nature Neuroscience , volume=. 2019 , publisher=

  49. [49]

    Dual-feature selectivity enables bidirectional coding in visual cortical neurons

    Franke, Katrin and Karantzas, Nikos and Willeke, Konstantin and Diamantaki, Maria and Ramakrishnan, Kandan and Bedel, Hasan Atakan and Elumalai, Pavithra and Restivo, Kelli and Fahey, Paul and Nealley, Cate and Shinn, Tori and Garcia, Gabrielle and Patel, Saumil and Ecker, Alexander and Walker, Edgar Y and Froudarakis, Emmanouil and Sanborn, Sophia and Si...

  50. [50]

    Soft Computing and Pattern Recognition (SoCPaR), 2014 6th International Conference of , pages=

    Fast classification of handwritten on-line Arabic characters , author=. Soft Computing and Pattern Recognition (SoCPaR), 2014 6th International Conference of , pages=. 2014 , organization=

  51. [51]

    arXiv preprint arXiv:1804.09028 , year=

    Estimate and Replace: A Novel Approach to Integrating Deep Neural Networks with Existing Applications , author=. arXiv preprint arXiv:1804.09028 , year=

  52. [52]

    TMLR , year=

    Interpreting Neurons in Deep Vision Networks with Language Models , author=. TMLR , year=

  53. [53]

    Advances in Neural Information Processing Systems , volume=

    Compositional explanations of neurons , author=. Advances in Neural Information Processing Systems , volume=

  54. [54]

    Proceedings of the IEEE conference on computer vision and pattern recognition , pages=

    Network dissection: Quantifying interpretability of deep visual representations , author=. Proceedings of the IEEE conference on computer vision and pattern recognition , pages=

  55. [55]

    Clip-dissect: Automatic description of neuron representations in deep vision networks,

    Clip-dissect: Automatic description of neuron representations in deep vision networks , author=. arXiv preprint arXiv:2204.10965 , year=

  56. [56]

    International conference on machine learning , pages=

    Learning transferable visual models from natural language supervision , author=. International conference on machine learning , pages=. 2021 , organization=

  57. [57]

    International Conference on Machine Learning , pages=

    Identifying interpretable subspaces in image representations , author=. International Conference on Machine Learning , pages=. 2023 , organization=

  58. [58]

    International Conference on Learning Representations , year=

    Natural language descriptions of deep visual features , author=. International Conference on Learning Representations , year=

  59. [59]

    Proceedings of the 6th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP , pages=

    Rigorously assessing natural language explanations of neurons , author=. Proceedings of the 6th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP , pages=

  60. [60]

    Advances in Neural Information Processing Systems , volume=

    Find: A function description benchmark for evaluating interpretability methods , author=. Advances in Neural Information Processing Systems , volume=

  61. [61]

    2009 IEEE conference on computer vision and pattern recognition , pages=

    Imagenet: A large-scale hierarchical image database , author=. 2009 IEEE conference on computer vision and pattern recognition , pages=. 2009 , organization=

  62. [62]

    Proceedings of the IEEE conference on computer vision and pattern recognition , pages=

    Understanding deep image representations by inverting them , author=. Proceedings of the IEEE conference on computer vision and pattern recognition , pages=

  63. [63]

    Distill , volume=

    Feature visualization , author=. Distill , volume=

  64. [64]

    European conference on computer vision , pages=

    Visualizing and understanding convolutional networks , author=. European conference on computer vision , pages=. 2014 , organization=

  65. [65]

    Proceedings of the IEEE conference on computer vision and pattern recognition , pages=

    Rich feature hierarchies for accurate object detection and semantic segmentation , author=. Proceedings of the IEEE conference on computer vision and pattern recognition , pages=

  66. [66]

    arXiv preprint arXiv:1506.02078 , year=

    Visualizing and understanding recurrent networks , author=. arXiv preprint arXiv:1506.02078 , year=

  67. [67]

    Proceedings of the AAAI Conference on Artificial Intelligence , volume=

    What is one grain of sand in the desert? analyzing individual neurons in deep nlp models , author=. Proceedings of the AAAI Conference on Artificial Intelligence , volume=

  68. [68]

    Distill , volume=

    Curve circuits , author=. Distill , volume=

  69. [69]

    Proceedings of the IEEE/CVF International Conference on Computer Vision , pages=

    Toward a visual concept vocabulary for gan latent space , author=. Proceedings of the IEEE/CVF International Conference on Computer Vision , pages=

  70. [70]

    Efros, and Jacob Steinhardt

    Interpreting clip's image representation via text-based decomposition , author=. arXiv preprint arXiv:2310.05916 , year=

  71. [71]

    Forty-first International Conference on Machine Learning , year=

    A multimodal automated interpretability agent , author=. Forty-first International Conference on Machine Learning , year=

  72. [72]

    Proceedings of the 40th International Conference on Machine Learning , pages =

    Identifying Interpretable Subspaces in Image Representations , author =. Proceedings of the 40th International Conference on Machine Learning , pages =. 2023 , editor =

  73. [73]

    International conference on machine learning , pages=

    Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav) , author=. International conference on machine learning , pages=. 2018 , organization=

  74. [74]

    arXiv preprint arXiv:1707.08139 , year=

    Analogs of linguistic structure in deep representations , author=. arXiv preprint arXiv:1707.08139 , year=

  75. [75]

    A tale of two tails: Preferred and anti-preferred natural stimuli in visual cortex

    Gondur, Rabia and Stan, Patricia L and Smith, Matthew A and Cowley, Benjamin R. A tale of two tails: Preferred and anti-preferred natural stimuli in visual cortex. The Fourteenth International Conference on Learning Representations

  76. [76]

    The importance of mixed selectivity in complex cognitive tasks

    Rigotti, Mattia and Barak, Omri and Warden, Melissa R and Wang, Xiao-Jing and Daw, Nathaniel D and Miller, Earl K and Fusi, Stefano. The importance of mixed selectivity in complex cognitive tasks. Nature

  77. [77]

    Local vs

    Thorpe, Simon. Local vs. Distributed Coding. Intellectica

  78. [78]

    International Conference on Learning Representations (ICLR) , year=

    Local vs distributed representations: What is the right basis for interpretability? , author=. International Conference on Learning Representations (ICLR) , year=. 2411.03993 , archivePrefix=

  79. [79]

    Deep learning-driven characterization of single cell tuning in primate visual area V4 supports topological organization

    Willeke, Konstantin F and Restivo, Kelli and Franke, Katrin and Nix, Arne F and Cadena, Santiago A and Shinn, Tori and Nealley, Cate and Rodriguez, Gabrielle and Patel, Saumil and Ecker, Alexander S and Sinz, Fabian H and Tolias, Andreas S. Deep learning-driven characterization of single cell tuning in primate visual area V4 supports topological organizat...

  80. [80]

    What's ``up'' with vision-language models? Investigating their struggle with spatial reasoning

    Kamath, Amita and Hessel, Jack and Chang, Kai-Wei. What's ``up'' with vision-language models? Investigating their struggle with spatial reasoning. arXiv [cs.CL]

Showing first 80 references.