A primary goal of systems neuroscience is to discover how ensembles of neurons transform inputs into goal-directed behavior, a process known as neural computation. A powerful framework for understanding neural computation uses neural dynamics – the rules that describe the temporal evolution of neural activity – to explain how goal-directed input-output transformations occur. As dynamical rules are not directly observable, we need computational models that can infer neural dynamics from recorded neural activity. We typically validate such models using synthetic datasets with known ground-truth dynamics, but unfortunately existing synthetic datasets don’t reflect fundamental features of neural computation and are thus poor proxies for neural systems. Further, the field lacks validated metrics for quantifying the accuracy of the dynamics inferred by models. The Computation-through-Dynamics Benchmark (CtDB) fills these critical gaps by providing: 1) synthetic datasets that reflect computational properties of biological neural circuits, 2) interpretable metrics for quantifying model performance, and 3) a standardized pipeline for training and evaluating models with or without known external inputs. In this manuscript, we demonstrate how CtDB can help guide the development, tuning, and troubleshooting of neural dynamics models. In summary, CtDB provides a critical platform for model developers to better understand and characterize neural computation through the lens of dynamics.Competing Interest StatementD.S. and C.P. are employees of Meta Platforms Reality Labs. Meta Platforms Reality Labs did not support this work, nor did it have any role in the study or any competing interests related to it.
2024
How Diffusion Models Learn to Factorize and Compose
Qiyao Liang, Ziming Liu, Mitchell Ostrow, and 1 more author
Neural Information Processing Systems, 2024
Does Maximizing Neural Regression Scores Teach Us About The Brain?
Rylan Schaeffer, Mikail Khona, Sarthak Chandra, and 3 more authors
UniReps: 2nd Edition of the Workshop on Unifying Representations in Neural Models, 2024
Delay Embedding Theory of Neural Sequence Models
Mitchell Ostrow, Adam Eisen, and Ila Fiete
International Conference on Machine Learning (ICML) workshop on Next Generation Sequence Modeling, 2024
To generate coherent responses, language models infer unobserved meaning from their input text sequence. One potential explanation for this capability arises from theories of delay embeddings in dynamical systems, which prove that unobserved variables can be recovered from the history of only a handful of observed variables. To test whether language models are effectively constructing delay embeddings, we measure the capacities of sequence models to reconstruct unobserved dynamics. We trained 1-layer transformer decoders and state-space sequence models on next-step prediction from noisy, partially-observed time series data. We found that each sequence layer can learn a viable embedding of the underlying system. However, state-space models have a stronger inductive bias than transformers-in particular, they more effectively reconstruct unobserved information at initialization, leading to more parameter-efficient models and lower error on dynamics tasks. Our work thus forges a novel connection between dynamical systems and deep learning sequence models via delay embedding theory.
@article{delay_embed,title={Delay Embedding Theory of Neural Sequence Models},author={Ostrow, Mitchell and Eisen, Adam and Fiete, Ila},year={2024},journal={International Conference on Machine Learning (ICML) workshop on Next Generation Sequence Modeling},url={https://arxiv.org/abs/2406.11993v1},topic={Understanding neural systems}}
2023
Beyond Geometry: Comparing the Temporal Structure of Computation in Neural Circuits with Dynamical Similarity Analysis
Mitchell Ostrow, Adam Eisen, Leo Kozachkov, and 1 more author
How can we tell whether two neural networks are utilizing the same internal processes for a particular computation? This question is pertinent for multiple subfields of both neuroscience and machine learning, including neuroAI, mechanistic interpretability, and brain-machine interfaces. Standard approaches for comparing neural networks focus on the spatial geometry of latent states. Yet in recurrent networks, computations are implemented at the level of neural dynamics, which do not have a simple one-to-one mapping with geometry. To bridge this gap, we introduce a novel similarity metric that compares two systems at the level of their dynamics. Our method incorporates two components: Using recent advances in data-driven dynamical systems theory, we learn a high-dimensional linear system that accurately captures core features of the original nonlinear dynamics. Next, we compare these linear approximations via a novel extension of Procrustes Analysis that accounts for how vector fields change under orthogonal transformation. Via four case studies, we demonstrate that our method effectively identifies and distinguishes dynamic structure in recurrent neural networks (RNNs), whereas geometric methods fall short. We additionally show that our method can distinguish learning rules in an unsupervised manner. Our method therefore opens the door to novel data-driven analyses of the temporal structure of neural computation, and to more rigorous testing of RNNs as models of the brain.
@article{ostrow2023geometry,title={Beyond Geometry: Comparing the Temporal Structure of Computation in Neural Circuits with Dynamical Similarity Analysis},author={Ostrow, Mitchell and Eisen, Adam and Kozachkov, Leo and Fiete, Ila},year={2023},url={https://arxiv.org/abs/2306.10168},journal={Neural Information Processing Systems},topic={ML methods for neural data analysis}}
Predictive Models are not Enough for Explanation-Seeking Curiosity: A Case Study.
Hokyung Sung, and Mitchell Ostrow
In Curiosity, Creativity and Complexity Conference, 2023
The use of an internal model to infer and predict others‘ mental states and actions, broadly referred to as Theory of Mind (ToM), is a fundamental aspect of human social intelligence. Nevertheless, it remains unknown how these models are used during social interactions, and how they help an agent generalize to new contexts. We investigated a putative neural mechanism of ToM in a recurrent circuit through the lens of an artificial neural network trained with reinforcement learning (RL) to play a competitive matching pennies game against many algorithmic opponents. The network showed near-optimal performance against unseen opponents, indicating that it had acquired the capacity to adapt against new strategies online. Analysis of recurrent states during play against out-of-training-distribution (OOD) opponents in relation to those of within-training-distribution (WD) opponents revealed two similarity-based mechanisms by which the network might generalize: mapping to a known strategy (template matching) or known opponent category (interpolation). Even when the network’s strategy cannot be explained by template-matching or interpolation, the recurrent activity fell upon the low-dimensional manifold of the WD neural activity, suggesting the contribution of prior experience with WD opponents. Furthermore, these states occupied low-density edges of the WD-manifold, suggesting that the network can extrapolate beyond any learned strategy or category. Our results suggest that a neural implementation for ToM may be a reservoir of learned representations that provide the capacity for generalization via flexible access and reuse of these stored features.
@article{MPRNN_article,title={Representational Geometry of Social Inference and Generalization in a Competitive Game},author={Ostrow, Mitchell and Yang, G.R. and Seo, H.},year={2022},journal={Robotics: Science and Systems},workshop={Social Intelligence in Humans and Robots},topic={Understanding neural systems}}
Neural Representations of Opponent Strategy Support the Adaptive Behavior of Recurrent Actor-Critics in a Competitive Game
Mitchell Ostrow, G.R. Yang, and H. Seo
In Computational and Systems Neuroscience (COSYNE), 2022
@misc{MPRNN_misc,title={Representational Geometry of Social Inference and Generalization in a Competitive Game},author={Ostrow, Mitchell and Yang, G.R. and Seo, H.},year={2022},howpublished={Robotics: Science and Systems (Spotlight Talk)},topic={Understanding neural systems}}
2021
Examining the Viability of Computational Psychiatry: Approaches into the Future
As modern medicine becomes increasingly personalized, psychiatry lags behind, using poorly-understood drugs and therapies to treat mental disorders. With the advent of methods that capture large quantities of data, such as genome-wide analyses or fMRI, machine learning (ML) approaches have become prominent in neuroscience. This is promising for studying the brain’s function, but perhaps more importantly, these techniques can potentially predict the onset of disorder and treatment response. Experimental approaches that use naive machine learning algorithms have dominated research in computational psychiatry over the past decade. In a critical review and analysis, I argue that biologically realistic approaches will be more effective in clinical practice, and research trends should reflect this. Hybrid models are considered, and a brief case study on major depressive disorder is presented. Finally, I propose a novel four-step approach for the future implementation of computational methods in psychiatric clinics.