Towards a cognitive science for neural networks
I’m a second year Computer Science PhD student @ Stanford, advised by Christopher Potts and Judith Fan. My primary research interests are deep neural network interpretability and multimodal reasoning, drawing inspiration from cognitive science. The questions I’m the most excited about lie at the intersection of vision, language, and thought. How do neural networks see the world, and why do they continue to struggle on some of the most fundamental tasks studied by cognitive scientists?
>> research directions <<
I study how neural networks represent and reason about the world, drawing on ideas from cognitive science. Current threads include:
- Symbols in neural networks. How do abstract concepts // symbolic structures emerge in DNNs? What mechanisms allow systems to move from continuous sensory input → discrete representations? Does a lack of symbolic structure explain persistent limitations of these models?
- Thinking across modalities. How do different modalities (e.g. vision, language) afford distinct kinds of reasoning? Is natural language really the best general-purpose medium for thought?
- Cognitively-inspired interpretability. How can we repurpose methodologies & frameworks from cognitive science & computational neuroscience to understand the internal workings of modern models? Conversely, what can probing them reveal about human intelligence?
- Representational alignment & universality. To what extent do intelligent systems converge on a shared representational space? What are the key representational differences between humans and machines, and how can we bridge them? How can we design benchmarks and evaluation protocols that enable meaningful human–machine comparisons?
{{ publications }}
- Alexa R. Tartaglini, Satchel Grant, Daniel Wurgaft, Christopher Potts & Judith E. Fan. Diagnosing Bottlenecks in Data Visualization Understanding by Vision-Language Models. Under review (2025)
- Satchel Grant & Alexa R. Tartaglini. Control and Predictivity in Neural Interpretability. NeurIPS MechInterp Workshop (2025)
- Michael A. Lepori, Alexa R. Tartaglini, Wai Keen Vong, Thomas Serre, Brenden M. Lake & Ellie Pavlick. Beyond the Doors of Perception: Vision Transformers Represent Relations Between Objects. NeurIPS (2024)
- Alexa R. Tartaglini, Sheridan Feucht, Michael A. Lepori, Wai Keen Vong, Charles Lovering, Brenden M. Lake & Ellie Pavlick. Deep Neural Networks Can Learn Generalizable Same-Different Visual Relations. Computational Cognitive Neuroscience Proceedings (2023)
- Alexa R. Tartaglini, Wai Keen Vong, Brenden M. Lake. A Developmentally-Inspired Examination of Shape versus Texture Bias in Machines. CogSci; oral (2022)
- Alexa R. Tartaglini, Wai Keen Vong & Brenden M. Lake. Modeling artificial category learning from pixels: Revisiting Shepard, Hovland, and Jenkins (1961) with deep neural networks. CogSci (2021)
:: contact ::
I’m always excited to collaborate or exchange ideas. If you'd like to talk about my work, your work, or anything really, please reach out!
alexart@stanford.edu
Stanford University
Stanford, CA 94305
Check out my social links above! Most active on Twitter/X.