Weekly Research Meetings

Research Meeting: Computational Feasibility of Artificial Human-Level Intelligence

Oct 29, 2024 - 4:00 pm
Venue:  Room 45-792 Speaker/s:  Eran Malach, Harvard University

Abstract: Modern machine learning models, in particular large language models, are approaching and even surpassing human-level performance at various benchmarks. In this talk, I will discuss the possibilities and barriers towards achieving human-level intelligence from a computational learning theory perspective. Specifically, I will talk about how auto-regressive next-token predictors can learn to solve computationally complex tasks. Additionally, I will discuss how generative models can “transcend” their training data, outperforming the experts that generate their data, with specific focus on learning to play chess from game transcripts.

Organizer:  Kathleen Sullivan Organizer Email:  cbmm-contact@mit.edu

Research Meeting: Lorenzo Rosasco

Sep 17, 2024 - 4:00 pm
Venue:  Room 45-792 Speaker/s:  Lorenzo Rosasco, Italian Institute of Technology (IIT), Università degli Studi di Genova

Abstract: Supervised learning is the problem of estimating a function from input and output samples. But how many samples are needed to achieve a prescribed accuracy?

This question can be answered only by restricting the class of problems—for example, considering functions that don’t vary much. But in even this case, we find that the number of needed samples depends exponentially on the dimensions of each input—the so-called curse of dimensionality.

Since neural nets seem to learn well with much less data, it is natural to postulate that the underlying problems (functions) have more structure beyond bounded variations.  The search for the right notion of “structure” has been quite elusive thus far, and I will discuss some recent results that emphasize the role of sparsity and compositions.

Bio: Lorenzo Rosasco is a professor at the University of Genova, a research affiliate at MIT, and a visiting scientist at the Italian Technological Institute (IIT). He is a founder and coordinator of the Machine Learning Genova center (MaLGa) and the Laboratory for Computational and Statistical Learning, focusing on the theory, algorithms, and applications of machine learning. He obtained his PhD in 2006 from the University of Genova and was a visiting student at the Center for Biological and Computational Learning at MIT, the Toyota Technological Institute at Chicago (TTI-Chicago), and the Johann Radon Institute for Computational and Applied Mathematics. From 2006 to 2013, he worked as a postdoc and research scientist at the Brain and Cognitive Sciences Department at MIT. He is a fellow at Ellis and serves as the co-director of the "Theory, Algorithms and Computations of Modern Learning Systems" program as well as the Ellis Genoa unit. Lorenzo has received several awards, including an ERC consolidator grant.

Organizer:  Kathleen Sullivan Organizer Email:  cbmm-contact@mit.edu

CBMM Research Meeting: Navigating the perceptual space with neural perturbations

Feb 27, 2024 - 3:00 pm
Venue:  McGovern Reading Room (46-5165) Speaker/s:  Arash Afraz Ph.D., Chief of unit on neurons, circuits and behavior, laboratory of neuropsychology, NIMH, NIH

Abstract: Local perturbation of neural activity in high-level visual cortical areas alters visual perception. Quantitative characterization of these perceptual alterations holds the key to understanding the mapping between patterns of neuronal activity and elements of perception. The complexity and subjectivity of these perceptual alterations makes them difficult to study. I introduce a new experimental approach, “Perceptography”, to develop “pictures” of the subjective experience induced by optogenetic cortical stimulation in the inferior temporal cortex of macaque monkeys. 

Bio: Dr. Arash Afraz received his MD from Tehran University of Medical Sciences in 2003. In 2005 he joined the Vision Science Laboratory at Harvard and studied spatial constraints of face recognition under the mentorship of Dr. Patrick Cavanagh. Dr. Afraz received his PhD in Psychology from Harvard University in 2009. Right after, he joined Dr. James DiCarlo’s group at MIT as a postdoctoral fellow to study the neural underpinnings of face and object recognition. Dr. Afraz started at NIMH as a principal investigator in 2017 to lead the unit on Neurons, Circuits and Behavior (Afraz group). Dr. Afraz’s group, Unit on Neurons, Circuits and Behavior, studies the neural mechanisms of visual object recognition. The research team is particularly interested in establishing causal links between the neural activity in the ventral stream of visual processing in the brain and object recognition behavior. The group combines visual psychophysics with conventional methods of single unit recording as well as microstimulation, drug microinjection and optogenetics to bridge the gap between the neural activity and visual perception.

Organizer:  Hector Penagos Organizer Email:  cbmm-contact@mit.edu

CBMM Research Meeting: A Neural Hypothesis for Language

Dec 7, 2023 - 2:30 pm
Venue:  McGovern Seminar Room (46-3189) Speaker/s:  Daniel Mitropolsky, Columbia University

Abstract: How do neurons, in their collective action, beget cognition, as well as intelligence and reasoning? As Richard Axel recently put it, we do not have a logic for the transformation of neural activity into thought and action; discerning this logic as the most important future direction of neuroscience. I will present a mathematical neural model of brain computation called NEMO, whose key ingredients are spiking neurons, random synapses and weights, local inhibition, and Hebbian plasticity (no backpropagation). Concepts are represented by interconnected co-firing assemblies of neurons that emerge organically from the dynamical system of its equations. It turns out it is possible to carry out complex operations on these concept representations, such as copying, merging, completion from small subsets, and sequence memorization. NEMO is a neuromorphic computational system that, because of its simplifying assumptions, can be efficiently simulated on modern hardware. I will present how to use NEMO to implement an efficient parser of a small but non-trivial subset of English, and a more recent model of the language organ in the baby brain that learns the meaning of words, and basic syntax, from whole sentences with grounded input. In addition to constituting hypotheses as to the logic of the brain, we will discuss how principles from these brain-like models might be used to improve AI, which, despite astounding recent progress, still lags behind humans in several key dimensions such as creativity, hard constraints, energy consumption.

Organizer:  Hector Penagos Organizer Email:  cbmm-contact@mit.edu

CBMM Weekly Research Meeting: Excitatory local lateral connectivity is sufficient for reproducing cortex-like topography

Nov 7, 2023 - 4:00 pm
Venue:  McGovern Reading Room (46-5165) Speaker/s:  Pouya Bashivan, McGill University

Abstract:

Across the primate neocortex, neurons that perform similar functions tend to be spatially grouped together. How such organization emerges and why have been debated extensively, with various models successfully replicating aspects of cortical topography using cost functions and learning rules designed to induce topographical structures. However, these models often compromise task learning capabilities and rely on strong assumptions about learning in neural circuits. I will introduce two new approaches for training topographically organized neural networks that substantially improve the trade-off between task performance and topography while also simplifying the assumptions about learning in neural circuits required to obtain brain-like topography. In particular, I will show that excitatory local lateral connectivity is sufficient for simulating cortex-like topographical organization without the need for any topography-promoting learning rules or objectives. I will also discuss the implications of this model for the link between topographical organization and robust representations. 

Bio:

Pouya Bashivan is an Assistant Professor at the Department of Physiology at McGill University, member of the Integrated Program in Neuroscience and an associate member of the Quebec AI Institute (MILA). Prior to joining McGill University, he was a postdoctoral fellow at MILA working with Drs. Irina Rish and Blake Richards. Prior to that he was a postdoctoral researcher at the Department of Brain and Cognitive Sciences and the McGovern Institute for Brain Research, MIT, working with Professor James DiCarlo. He received his PhD in computer engineering from the University of Memphis in 2016. Before that, Pouya studied control engineering and earned a B.Sc. and a M.Sc. degree in electrical and control engineering from KNT University (Tehran, Iran).

The goal of research in Bashivan lab is to develop neural network models that leverage memory to solve complex tasks. While we often rely on task-performance measures to find improved neural network models and learning algorithms, we also use neural and behavioral measurements from humans and other animal brains to evaluate the similarity of these models to biologically evolved brains. We believe that these additional constraints could expedite the progress towards engineering a human-level artificially-intelligent agent.

Organizer:  Hector Penagos Organizer Email:  cbmm-contact@mit.edu

CBMM Research Meeting: Panel - Associative Memories and Deep Learning

Sep 26, 2023 - 4:00 pm
Venue:  McGovern Reading Room (46-5165) Speaker/s:  Moderator: Ila Fiete; Panelists: Akshay Rangamani, Tomer Galanti, Sugandha Sharma and Sarthak Chandra

Abstract: 

This discussion will feature panelists from the Poggio and Fiete labs describing ongoing work on deep classifiers and hippocampal-inspired networks, respectively. Following brief presentations by each panelist, Ila Fiete will steer a discussion to explore strengths, limitations and neural plausibility of each approach.

Sugandha Sharma and Sarthak Chandra (Fiete lab).

Hippocampal circuits in the brain enable two distinct cognitive functions: construction of spatial maps for navigation and storage of sequential episodic memories. This dual role of the hippocampus remains an enduring enigma. While there have been advances in modeling the spatial representation properties of the hippocampus, we lack good models of its role in episodic memory. Here we present a neocortical-entorhinal-hippocampal network model that exhibits high-capacity general episodic memory and spatial memory without the memory cliff of existing neural memory models. Instead, the circuit exhibits a graceful tradeoff between number of stored items and detail, achieved by factorizing content storage from the dynamics of generating error-correcting stable states. The exponentially large space avoids catastrophic forgetting. Next, we show that pre-structured representations are an essential feature for constructing episodic memory: unlike existing episodic memory models, they enable high-capacity memorization of sequences by abstracting the chaining problem into one of learning transitions within a rigid low-dimensional grid cell scaffold. Finally, we show that previously learned spatial sequences in the form of location-landmark associations can themselves be re-usably leveraged as robust scaffolds and associated with neocortical inputs for a high-fidelity one-shot memory, providing the first circuit model of the ``memory palaces'' used in the striking feats of memory athletes.

Akshay Rangamani and Tomer Galanti (Poggio lab).

We conduct an empirical study of the feature learning process in deep classifiers. Recent research has identified a training phenomenon called Neural Collapse (NC), in which the top- layer feature embeddings of samples from the same class tend to concentrate around their means, and the top layer’s weights align with those features. Our study aims to investigate if these properties extend to intermediate layers. We empirically study the evolution of the covariance and mean of representations across different layers and show that as we move deeper into a trained neural network, the within-class covariance decreases relative to the between-class covariance. Additionally, we find that in the top layers, where the between-class covariance is dominant, the subspace spanned by the class means aligns with the subspace spanned by the most significant singular vector components of the weight matrix in the corresponding layer. Finally, we discuss the relationship between NC and Associative Memories.

Organizer:  Hector Penagos Organizer Email:  cbmm-contact@mit.edu

CBMM Research Meeting: Using Embodied AI to help answer “why” questions in systems neuroscience

Sep 19, 2023 - 4:00 pm
Venue:  MIBR Reading Room 46-5165 Speaker/s:  Aran Nayebi, ICoN Postdoctoral Fellow at MIT

Abstract:

Deep neural networks trained on high-variation tasks ("goals”) have had immense success as predictive models of the human and non-human primate visual pathways. More specifically, a positive relationship has been observed between model performance on ImageNet categorization and neural predictivity. Past a point, however, improved categorization performance on ImageNet does not yield improved neural predictivity, even between very different architectures. In this talk, I will present two case studies in both rodents and primates, that demonstrate a more general correspondence between self-supervised learning of visual representations relevant to high-dimensional embodied control and increased gains in neural predictivity.

In the first study, we develop the (currently) most precise model of the mouse visual system, and show that self-supervised, contrastive algorithms outperform supervised approaches in capturing neural response variance across visual areas. By “implanting” these visual networks into a biomechanically-realistic rodent body to navigate to rewards in a novel maze environment, we observe that the artificial rodent with a contrastively-optimized visual system is able to obtain more reward across episodes compared to its supervised counterpart. The second case study examines mental simulations in primates, wherein we show that self-supervised video foundation models that predict the future state of their environment in latent spaces that can support a wide range of sensorimotor tasks, align most closely with human error patterns and macaque frontal cortex neural dynamics. Taken together, our findings suggest that self-supervised learning of visual representations that are reusable for downstream Embodied AI tasks may be a promising way forward to study the evolutionary constraints of neural circuits in multiple species.

Organizer:  Hector Penagos Organizer Email:  cbmm-contact@mit.edu

CBMM Research Meeting: Do we really need to train by minimizing a loss functional?

Mar 28, 2023 - 4:00 pm
Venue:  McGovern Reading Room (46-5165) Speaker/s:  H. N. Mhaskar - Claremont Graduate University, Claremont.

The fundamental problem of machine learning is often formulated as the problem of function approximation. For example, we have data of the form {(xj,yj)}, where yj is the class label for xj, and we want to approximate the class label as a function of the input x. The standard way for this approximation is to minimize a loss functional, usually with some regularization. Surprisingly, even though the problem is posed as a problem of function approxi- mation, approximation theory has played only a marginal role in this theory. We describe our efforts to explore why this might be the case, and also to develop approximation theory/harmonic analysis tools more meaningfully and directly applicable to machine learning. We also argue that classification problems are better treated as generalized signal separation problems rather than function approximation problems.

Organizer:  Hector Penagos Organizer Email:  cbmm-contact@mit.edu

CBMM Research Meeting: Towards data-driven modeling in large-scale naturalistic neuroscience

Feb 28, 2023 - 4:00 pm
Venue:  McGovern Reading Room (46-5165) Speaker/s:  Meenakshi Khosla (Kanwisher lab postdoc)

Abstract: Neuroscience is currently undergoing an explosion in the availability of large-scale brain activity data, so the major challenge no longer lies in data collection, but also in deriving understanding from this abundant stream of complex, high-dimensional, noisy data with methods that fully leverage its potential. How can we understand neural representations and infer computational principles from large-scale brain activity data directly?

In this talk, I will present several lines of previous research aimed at tackling this question. First, I will present a line of data-driven modeling work that revealed the representational structure in the high-level visual cortex and led to the discovery of a neural population selectively responsive to images of food. Second, I will present a modeling framework, called response-optimization, for inferring computations directly from brain activity data with minimal apriori hypotheses. Here, we trained artificial neural network (ANN) models directly to predict the brain activity related to viewing natural images. We then developed techniques for interpreting the networks and characterizing the emergent functional capabilities of these brain response-optimized networks. This work highlights how models trained to capture human brain activity can spontaneously recapitulate human-like behavior. Finally, I will present my work on developing neural network models of brain responses across wide-spread cortical regions to dynamic, multi-modal stimuli like movies, with an integrated modeling approach that captured visual attention, multi-sensory auditory-visual interactions and temporal context.

This will be an in person only meeting.

Organizer:  Hector Penagos Organizer Email:  cbmm-contact@mit.edu

Pages

Subscribe to Weekly Research Meetings