What powerful new bots like ChatGPT tell us about intelligence and the human brain

This story originally appeared in the Spring 2023 issue of BrainScan.

___

Artificial intelligence seems to have gotten a lot smarter recently. AI technologies are increasingly integrated into our lives — improving our weather forecasts, finding efficient routes through traffic, personalizing the ads we see and our experiences with social media.

Watercolor image of a robot with a human brain, created using the AI system DALL*E2.

But with the debut of powerful new chatbots like ChatGPT, millions of people have begun interacting with AI tools that seem convincingly human-like. Neuroscientists are taking note — and beginning to dig into what these tools tell us about intelligence and the human brain.

The essence of human intelligence is hard to pin down, let alone engineer. McGovern scientists say there are many kinds of intelligence, and as humans, we call on many different kinds of knowledge and ways of thinking. ChatGPT’s ability to carry on natural conversations with its users has led some to speculate the computer model is sentient, but McGovern neuroscientists insist that the AI technology cannot think for itself.

Still, they say, the field may have reached a turning point.

“I still don’t believe that we can make something that is indistinguishable from a human. I think we’re a long way from that. But for the first time in my life I think there is a small, nonzero chance that it may happen in the next year,” says McGovern founding member Tomaso Poggio, who has studied both human intelligence and machine learning for more than 40 years.

Different sort of intelligence

Developed by the company OpenAI, ChatGPT is an example of a deep neural network, a type of machine learning system that has made its way into virtually every aspect of science and technology. These models learn to perform various tasks by identifying patterns in large datasets. ChatGPT works by scouring texts and detecting and replicating the ways language is used. Drawing on language patterns it finds across the internet, ChatGPT can design you a meal plan, teach you about rocket science, or write a high school-level essay about Mark Twain. With all of the internet as a training tool, models like this have gotten so good at what they do, they can seem all-knowing.

“Engineers have been inventing some of these forms of intelligence since the beginning of the computers. ChatGPT is one. But it is very far from human intelligence.” – Tomaso Poggio

Nonetheless, language models have a restricted skill set. Play with ChatGPT long enough and it will surely give you some wrong information, even if its fluency makes its words deceptively convincing. “These models don’t know about the world, they don’t know about other people’s mental states, they don’t know how things are beyond whatever they can gather from how words go together,” says Postdoctoral Associate Anna Ivanova, who works with McGovern Investigators Evelina Fedorenko and Nancy Kanwisher as well as Jacob Andreas in MIT’s Computer Science and Artificial Intelligence Laboratory.

Such a model, the researchers say, cannot replicate the complex information processing that happens in the human brain. That doesn’t mean language models can’t be intelligent — but theirs is a different sort of intelligence than our own. “I think that there is an infinite number of different forms of intelligence,” says Poggio. “Engineers have been inventing some of these forms of intelligence since the beginning of the computers. ChatGPT is one. But it is very far from human intelligence.”

Under the hood

Just as there are many forms of intelligence, there are also many types of deep learning models — and McGovern researchers are studying the internals of these models to better understand the human brain.

A watercolor painting of a robot generated by DALL*E2.

“These AI models are, in a way, computational hypotheses for what the brain is doing,” Kanwisher says. “Up until a few years ago, we didn’t really have complete computational models of what might be going on in language processing or vision. Once you have a way of generating actual precise models and testing them against real data, you’re kind of off and running in a way that we weren’t ten years ago.”

Artificial neural networks echo the design of the brain in that they are made of densely interconnected networks of simple units that organize themselves — but Poggio says it’s not yet entirely clear how they work.

No one expects that brains and machines will work in exactly the same ways, though some types of deep learning models are more humanlike in their internals than others. For example, a computer vision model developed by McGovern Investigator James DiCarlo responds to images in ways that closely parallel the activity in the visual cortex of animals who are seeing the same thing. DiCarlo’s team can even use their model’s predictions to create an image that will activate specific neurons in an animal’s brain.

“We shouldn’t just automatically assume that if we trained a deep network on a task, that it’s going to look like the brain.” – Ila Fiete

Still, there is reason to be cautious in interpreting what artificial neural networks tell us about biology. “We shouldn’t just automatically assume that if we trained a deep network on a task, that it’s going to look like the brain,” says McGovern Associate Investigator Ila Fiete. Fiete acknowledges that it’s tempting to think of neural networks as models of the brain itself due to their architectural similarities — but she says so far, that idea remains largely untested.

McGovern Institute Associate Investigator Ila Fiete builds theoretical models of the brain. Photo: Caitlin Cunningham

She and her colleagues recently experimented with neural networks that estimate an object’s position in space by integrating information about its changing velocity.

In the brain, specialized neurons known as grid cells carry out this calculation, keeping us aware of where we are as we move through the world. Other researchers had reported that not only can neural networks do this successfully, those that do include components that behave remarkably like grid cells. They had argued that the need to do this kind of path integration must be the reason our brains have grid cells — but Fiete’s team found that artificial networks don’t need to mimic the brain to accomplish this brain-like task. They found that many neural networks can solve the same problem without grid cell-like elements.

One way investigators might generate deep learning models that do work like the brain is to give them a problem that is so complex that there is only one way of solving it, Fiete says.

Language, she acknowledges, might be that complex.

“This is clearly an example of a super-rich task,” she says. “I think on that front, there is a hope that they’re solving such an incredibly difficult task that maybe there is a sense in which they mirror the brain.”

Language parallels

In Fedorenko’s lab, where researchers are focused on identifying and understanding the brain’s language processing circuitry, they have found that some language models do, in fact, mimic certain aspects of human language processing. Many of the most effective models are trained to do a single task: make predictions about word use. That’s what your phone is doing when it suggests words for your text message as you type. Models that are good at this, it turns out, can apply this skill to carrying on conversations, composing essays, and using language in other useful ways. Neuroscientists have found evidence that humans, too, rely on word prediction as a part of language processing.

Fedorenko and her team compared the activity of language models to the brain activity of people as they read or listened to words, sentences, and stories, and found that some models were a better match to human neural responses than others. “The models that do better on this relatively unsophisticated task — just guess what comes next — also do better at capturing human neural responses,” Fedorenko says.

A watercolor painting of a language model, generated by DALL*E2.

It’s a compelling parallel, suggesting computational models and the human brain may have arrived at a similar solution to a problem, even in the face of the biological constraints that have shaped the latter. For Fedorenko and her team, it’s sparked new ideas that they will explore, in part, by modifying existing language models — possibly to more closely mimic the brain.

With so much still unknown about how both human and artificial neural networks learn, Fedorenko says it’s hard to predict what it will take to make language models work and behave more like the human brain. One possibility they are exploring is training a model in a way that more closely mirrors the way children learn language early in life.

Another question, she says, is whether language models might behave more like humans if they had a more limited recall of their own conversations. “All of the state-of-the-art language models keep track of really, really long linguistic contexts. Humans don’t do that,” she says.

Chatbots can retain long strings of dialogue, using those words to tailor their responses as a conversation progresses, she explains. Humans, on the other hand, must cope with a more limited memory. While we can keep track of information as it is conveyed, we only store a string of about eight words as we listen or read. “We get linguistic input, we crunch it up, we extract some kind of meaning representation, presumably in some more abstract format, and then we discard the exact linguistic stream because we don’t need it anymore,” Fedorenko explains.

Language models aren’t able to fill in gaps in conversation with their own knowledge and awareness in the same way a person can, Ivanova adds. “That’s why so far they have to keep track of every single input word,” she says. “If we want a model that models specifically the [human] language network, we don’t need to have this large context window. It would be very cool to train those models on those short windows of context and see if it’s more similar to the language network.”

Multimodal intelligence

Despite these parallels, Fedorenko’s lab has also shown that there are plenty of things language circuits do not do. The brain calls on other circuits to solve math problems, write computer code, and carry out myriad other cognitive processes. Their work makes it clear that in the brain, language and thought are not the same.

That’s borne out by what cognitive neuroscientists like Kanwisher have learned about the functional organization of the human brain, where circuit components are dedicated to surprisingly specific tasks, from language processing to face recognition.

“The upshot of cognitive neuroscience over the last 25 years is that the human brain really has quite a degree of modular organization,” Kanwisher says. “You can look at the brain and say, ‘what does it tell us about the nature of intelligence?’ Well, intelligence is made up of a whole bunch of things.”

In generating this image from the text prompt, “a watercolor painting of a woman looking in a mirror and seeing a robot,” DALL*E2 incorrectly placed the woman (not the robot) in the mirror, highlighting one of the weaknesses of current deep learning models.

In January, Fedorenko, Kanwisher, Ivanova, and colleagues shared an extensive analysis of the capabilities of large language models. After assessing models’ performance on various language-related tasks, they found that despite their mastery of linguistic rules and patterns, such models don’t do a good job using language in real-world situations. From a neuroscience perspective, that kind of functional competence is distinct from formal language competence, calling on not just language-processing circuits but also parts of the brain that store knowledge of the world, reason, and interpret social interactions.

Language is a powerful tool for understanding the world, they say, but it has limits.

“If you train on language prediction alone, you can learn to mimic certain aspects of thinking,” Ivanova says. “But it’s not enough. You need a multimodal system to carry out truly intelligent behavior.”

The team concluded that while AI language models do a very good job using language, they are incomplete models of human thought. For machines to truly think like humans, Ivanova says, they will need a combination of different neural nets all working together, in the same way different networks in the human brain work together to achieve complex cognitive tasks in the real world.

It remains to be seen whether such models would excel in the tech world, but they could prove valuable for revealing insights into human cognition — perhaps in ways that will inform engineers as they strive to build systems that better replicate human intelligence.

Yang Dan named winner of the 2023 Scolnick Prize in Neuroscience

The McGovern Institute announced today that the 2023 Edward M. Scolnick Prize in Neuroscience will be awarded to neurobiologist Yang Dan. Dan holds the Nan Fung Life Sciences Chancellor’s Chair in Neuroscience at the University of California, Berkeley, and has been a Howard Hughes Investigator since 2008. The Scolnick Prize is awarded annually by the McGovern Institute for outstanding achievements in neuroscience.

“Yang Dan’s systems-level experimentation to identify the cell types and circuits that control sleep cycles represents the highest level of neuroscience research,” says Robert Desimone, McGovern Institute director and chair of the selection committee. “Her work has defined precise mechanisms for how motor behaviors are suppressed during sleep and activated during arousal, with potential implications for the design of more targeted sedatives and the treatment of sleep disorders.”

Significance of sleep

Dan received a BS in Physics in 1988 from Peking University in China. She then moved to the US to obtain her PhD in neurobiology from Columbia University, in 1994, under the mentorship of Professor Mu-Ming Poo. Her doctoral research focused on mechanisms of plasticity at the neuromuscular synapse and was published in Science, Nature, and Neuron. During this time, she showed that the quantal release of neurotransmitters is not unique to neuronal cell types and, as one example, that retrograde signaling from muscle cells regulates the synaptic strength of the neuromuscular junction. For her postdoctoral training, Dan joined Clay Reid’s lab at The Rockefeller University and then accompanied Reid’s move to Harvard Medical School a short time later. Within just over two years, Yang had collected and analyzed neuronal recording data to support and develop key computational models of visual information coding – her two papers describing this work have been cited, together, over 900 times.

Yang Dan started her own laboratory in January 1997 when she joined the faculty of UC Berkeley’s Department of Molecular and Cell Biology as an assistant professor; she became a full professor in 2005. Dan’s lab became known for discoveries of how sensory inputs, especially visual inputs, are processed by the brain to influence behavior. Using electrophysiological recordings in model animals and computational analyses, her group worked out rules for how synaptic plasticity and neural connectivity, at the microcircuit and brain-wide level, contribute to learning and goal-directed behaviors.

Sleep recordings in various animal models and humans, shown in a research review by Yang Dan (2019 Annual Review of Neuroscience). (a) In nonmammalian animals such as jellyfish, Caenorhabditis elegans, Drosophila, and zebrafish, locomotor assay is used to measure sleep. (b) Examples of mouse EEG and EMG recordings during wakefulness and NREM and REM sleep. (c) Example polysomnography recordings from a healthy human subject during wakefulness and NREM (stage 3) and phasic REM sleep.

The Dan lab carved out a new research direction upon their discovery of mechanisms controlling rapid eye movement (REM) sleep, a state in which the brain is active and neuroplastic despite minimal sensory input. In their 2015 Nature paper, Dan’s group showed that, in mice, optogenetic activation of inhibitory neurons that project forward from the brainstem to the middle of the brain can instantaneously induce REM sleep. Since then, the Dan lab has published nearly a dozen primary research papers on the sleep-wake cycle that capitalize on the latest neural engineering techniques to record and control specific cell types and circuits in the brain. Most recently, she reported the discovery of neurons in the midbrain that receive wide-ranging inputs to coordinate active suppression of movement during REM and non-REM sleep with the release of movement during arousal. This circuit is key to the ability, known to exist in most animals, to experience sleep and even vivid dreaming without acting out. Dan’s discoveries are paving the way to a holistic understanding, from the molecular to macrocircuit levels, of how our bodies regulate sleep, an evolutionarily conserved behavior that is essential for survival.

Awards and honors

Dan was appointed as a Howard Hughes Medical Institute Investigator in 2008 and elected to the US National Academy of Sciences in 2018. She was awarded the Li Ka Shing Women in Science Award in 2007 and a Research Award for Innovation in Neuroscience from the Society for Neuroscience in 2009. She teaches summer courses at institutes around the world and has mentored 16 graduate students and 27 postdoctoral researchers, 25 of whom now run their own independent laboratories. Currently, Dan serves as an editorial board member on top-ranked science journals including Cell, Neuron, PNAS, and Current Opinion in Neurobiology.

Yang Dan will be awarded the Scolnick Prize on Wednesday, June 7, 2023. At 4:00 pm on that day, she will deliver a lecture titled “The how and why of sleep,” to be followed by a reception at the McGovern Institute, 43 Vassar Street (building 46, room 3002) in Cambridge. The event is free and open to the public.

 

 

Partnership with MIT Museum explores relationship between neuroscience and society

What does a healthy relationship between neuroscience and society look like? How do we set the conditions for that relationship to flourish? Researchers and staff at the McGovern Institute and the MIT Museum have been exploring these questions with a five-month planning grant from the Dana Foundation.

Between October 2022 and March 2023, the team tested the potential for an MIT Center for Neuroscience and Society through a series of MIT-sponsored events that were attended by students and faculty of nearby Cambridge Public Schools. The goal of the project was to learn more about what happens when the distinct fields of neuroscience, ethics, and public engagement are brought together to work side-by-side.

Researchers assist volunteer in mock MRI scanner
Gabrieli lab members Sadie Zacharek (left) and Shruti Nishith (right) demonstrate how the MRI mock scanner works with a student volunteer from the Cambridge Public Schools. Photo: Emma Skakel, MIT Museum

Middle schoolers visit McGovern

Over four days in February, more than 90 sixth graders from Rindge Avenue Upper Campus (RAUC) in Cambridge, Massachusetts, visited the McGovern Institute and participated in hands-on experiments and discussions about the ethical, legal, and social implications of neuroscience research. RAUC is one of four middle schools in the city of Cambridge with an economically, racially, and culturally diverse student population. The middle schoolers interacted with an MIT team led by McGovern Scientific Advisor Jill R. Crittenden, including seventeen McGovern neuroscientists, three MIT Museum outreach coordinators, and neuroethicist Stephanie Bird, a member of the Dana Foundation planning grant team.

“It is probably the only time in my life I will see a real human brain.” – RAUC student

The students participated in nine activities each day, including trials of brain-machine interfaces, close-up examinations of preserved human brains, a tour of McGovern’s imaging center in which students watched as their teacher’s brain was scanned, and a visit to the MIT Museum’s interactive Artificial Intelligence Gallery.

Imagine-IT, a brain-machine interface designed by a team of middle school students during a visit to the McGovern Institute.

To close out their visit, students worked in groups alongside experts to invent brain-computer interfaces designed to improve or enhance human abilities. At each step, students were introduced to ethical considerations through consent forms, questions regarding the use of animal and human brains, and the possible impacts of their own designs on individuals and society.

“I admit that prior to these four days, I would’ve been indifferent to the inclusion of children’s voices in a discussion about technically complex ethical questions, simply because they have not yet had any opportunity to really understand how these technologies work,” says one researcher involved in the visit. “But hearing the students’ questions and ideas has changed my perspective. I now believe it is critically important that all age groups be given a voice when discussing socially relevant issues, such as the ethics of brain computer interfaces or artificial intelligence.”

 

For more information on the proposed MIT Center for Neuroscience and Society, visit the MIT Museum website.

2023 MacVicar Faculty Fellows named

The Office of the Vice Chancellor and the Registrar’s Office have announced this year’s Margaret MacVicar Faculty Fellows: professor of brain and cognitive sciences John Gabrieli, associate professor of literature Marah Gubar, professor of biology Adam C. Martin, and associate professor of architecture Lawrence “Larry” Sass.

For more than 30 years, the MacVicar Faculty Fellows Program has recognized exemplary and sustained contributions to undergraduate education at MIT. The program is named in honor of Margaret MacVicar, the first dean for undergraduate education and founder of the Undergraduate Research Opportunities Program (UROP). New fellows are chosen every year through a competitive nomination process that includes submission of letters of support from colleagues, students, and alumni; review by an advisory committee led by the vice chancellor; and a final selection by the provost. Fellows are appointed to a 10-year term and receive $10,000 per year of discretionary funds.

Gabrieli, Gubar, Martin, and Sass join an elite group of more than 130 scholars from across the Institute who are committed to curricular innovation, excellence in teaching, and supporting students both in and out of the classroom.

John Gabrieli

“When I learned of this wonderful honor, I felt gratitude — for how MIT values teaching and learning, how my faculty colleagues bring such passion to their teaching, and how the students have such great curiosity for learning,” says new MacVicar Fellow John Gabrieli.

Gabrieli PhD ’87 received a bachelor’s degree in English from Yale University and his PhD in behavioral neuroscience from MIT. He is the Grover M. Hermann Professor in the Department of Brain and Cognitive sciences. Gabrieli is also an investigator in the McGovern Institute for Brain Research and the founding director of the MIT Integrated Learning Initiative (MITili). He holds appointments in the Department of Psychiatry at Massachusetts General Hospital and the Harvard Graduate School of Education, and studies the organization of memory, thought, and emotion in the human brain.

He joined Course 9 as a professor in 2005 and since then, he has taught over 3,000 undergraduates through the department’s introductory course, 9.00 (Introduction to Psychological Science). Gabrieli was recognized with departmental awards for excellence in teaching in 2009, 2012, and 2015. Highly sought after by undergraduate researchers, the Gabrieli Laboratory (GabLab) hosts five to 10 UROPs each year.

A unique element of Gabrieli’s classes is his passionate, hands-on teaching style and his use of interactive demonstrations, such as optical illusions and personality tests, to help students grasp some of the most fundamental topics in psychology.

His former teaching assistant Daniel Montgomery ’22 writes, “I was impressed by his enthusiasm and ability to keep students engaged throughout the lectures … John clearly has a desire to help students become excited about the material he’s teaching.”

Senior Elizabeth Carbonell agrees: “The excitement professor Gabrieli brought to lectures by starting with music every time made the classroom an enjoyable atmosphere conducive to learning … he always found a way to make every lecture relatable to the students, teaching psychological concepts that would shine a light on our own human emotions.”

Lecturer and 9.00 course coordinator Laura Frawley says, “John constantly innovates … He uses research-based learning techniques in his class, including blended learning, active learning, and retrieval practice.” His findings on blended learning resulted in two MITx offerings including 9.00x (Learning and Memory), which utilizes a nontraditional approach to assignments and exams to improve how students retrieve and remember information.

In addition, he is known for being a devoted teacher who believes in caring for the student as a whole. Through MITili’s Mental Wellness Initiative, Gabrieli, along with a compassionate team of faculty and staff, are working to better understand how mental health conditions impact learning.

Associate department head and associate professor of brain and cognitive sciences Josh McDermott calls him “an exceptional educator who has left his mark on generations of MIT undergraduate students with his captivating, innovative, and thoughtful approach to teaching.”

Mariana Gomez de Campo ’20 concurs: “There are certain professors that make their mark on students’ lives; professor Gabrieli permanently altered the course of mine.”

Laura Schulz, MacVicar Fellow and associate department head of brain and cognitive sciences, remarks, “His approach is visionary … John’s manner with students is unfailingly gracious … he hastens to remind them that they are as good as it gets, the smartest and brightest of their generation … it is the kind of warm, welcoming, inclusive approach to teaching that subtly but effectively reminds students that they belong here at MIT … It is little wonder that they love him.”

Marah Gubar

Marah Gubar joined MIT as an associate professor of literature in 2014. She received her BA in English literature from the University of Michigan at Ann Arbor and a PhD from Princeton University. Gubar taught in the English department at the University of Pittsburgh and served as director of the Children’s Literature Program. She received MIT’s James A. and Ruth Levitan Teaching Award in 2019 and the Teaching with Digital Technology Award in 2020.

Gubar’s research focuses on children’s literature, history of children’s theater, performance, and 19th- and 20th-century representations of childhood. Her research and pedagogies underscore the importance of integrated learning.

Colleagues at MIT note her efficacy in introducing new concepts and new subjects into the literature curriculum during her tenure as curricular chair. Gubar set the stage for wide-ranging curricular improvements, resulting in a host of literature subjects on interrelated topics within and across disciplines.

Gubar teaches several classes, including 21L.452 (Literature and Philosophy) and 21L.500 (How We Got to Hamilton). Her lectures provide uniquely enriching learning experiences in which her students are encouraged to dive into literary texts; craft thoughtful, persuasive arguments; and engage in lively intellectual debate.

Gubar encourages others to bring fresh ideas and think outside the box. For example, her seminar on “Hamilton” challenges students to recontextualize the hip-hop musical in several intellectual traditions. Professor Eric Klopfer, head of the Comparative Media Studies Program/Writing and interim head of literature, calls Gubar “a thoughtful, caring instructor, and course designer … She thinks critically about whose story is being told and by whom.”

MacVicar Fellow and professor of literature Stephen Tapscott praises her experimentation, abstract thinking, and storytelling: “Professor Gubar’s ability to frame intellectual questions in terms of problems, developments, and performance is an important dimension of the genius of her teaching.”

“Marah is hands-down the most enthusiastic, effective, and engaged professor I had the pleasure of learning from at MIT,” writes one student. “She’s one of the few instructors I’ve had who never feels the need to reassert her place in the didactic hierarchy, but approaches her students as intellectual equals.”

Tapscott continues, “She welcomes participation in ways that enrich the conversation, open new modes of communication, and empower students as autonomous literary critics. In professor Gubar’s classroom we learn by doing … and that progress also includes ‘doing’ textual analysis, cultural history, and abstract literary theory.”

Gubar is also a committed mentor and student testimonials highlight her supportive approach. One of her former students remarked that Gubar “has a strong drive to be inclusive, and truly cares about ‘getting it right’ … her passion for literature and teaching, together with her drive for inclusivity, her ability to take accountability, and her compassion and empathy for her students, make [her] a truly remarkable teacher.”

On receiving this award Marah Gubar writes, “The best word I can think of to describe how I reacted to hearing that I had received this very overwhelming honor is ‘plotzing.’ The Yiddish verb ‘to plotz’ literally means to crack, burst, or collapse, so that captures how undone I was. I started to cry, because it suddenly struck me how much joy my father, Edward Gubar, would have taken in this amazing news. He was a teacher, too, and he died during the first phase of this terrible pandemic that we’re still struggling to get through.”

Adam C. Martin

Adam C. Martin is a professor and undergraduate officer in the Department of Biology. He studies the molecular mechanisms that underlie tissue form and function. His research interests include gastrulation, embryotic development, cytoskeletal dynamics, and the coordination of cellular behavior. Martin received his PhD from the University of California at Berkeley and his BS in biology (genetics) from Cornell University. Martin joined the Course 7 faculty in 2011.

“I am overwhelmed with gratitude knowing that this has come from our students. The fact that they spent time to contribute to a nomination is incredibly meaningful to me,” says Martin. “I want to also thank all of my faculty colleagues with whom I have taught, appreciate, and learned immensely from over the past 12 years. I am a better teacher because of them and inspired by their dedication.”

He is committed to undergraduate education, teaching several key department offerings including 7.06 (Cell Biology), 7.016 (Introductory Biology), 7.002 (Fundamentals of Experimental Molecular Biology), and 7.102 (Introduction to Molecular Biology Techniques).

Martin’s style combines academic and scientific expertise with creative elements like props and demonstrations. His “energy and passion for the material” is obvious, writes Iain Cheeseman, associate department head and the Herman and Margaret Sokol Professor of Biology. “In addition to creating engaging lectures, Adam went beyond the standard classroom requirements to develop videos and animations (in collaboration with the Biology MITx team) to illustrate core cell biological approaches and concepts.”

What sets Martin apart is his connection with students, his positive spirit, and his welcoming demeanor. Apolonia Gardner ’22 reflects on the way he helped her outside of class through his running group, which connects younger students with seniors in his lab. “Professor Martin was literally committed to ‘going the extra mile’ by inviting his students to join him on runs around the Charles River on Friday afternoons,” she says.

Amy Keating, department head and Jay A. Stein professor of biology, and professor of biological engineering, goes on to praise Martin’s ability to attract students to Course 7 and guide them through their educational experience in his role as the director of undergraduate studies. “He hosts social events, presides at our undergraduate research symposium and the department’s undergraduate graduation and awards banquet, and works with the Biology Undergraduate Student Association,” she says.

As undergraduate officer, Martin is involved in both advising and curriculum building. He mentors UROP students, serves as a first-year advisor, and is a current member of MIT’s Committee on the Undergraduate Program (CUP).

Martin also brings a commitment to diversity, equity, and inclusion (DEI) as evidenced by his creation of a DEI journal club in his lab so that students have a dedicated space to discuss issues and challenges. Course 7 DEI officer Hallie Dowling-Huppert writes that Martin “thinks deeply about how DEI efforts are created to ensure that department members receive the maximum benefit. Adam considers all perspectives when making decisions, and is extremely empathetic and caring towards his students.”

“He makes our world so much better,” Keating observes. “Adam is a gem.”

Lawrence “Larry” Sass

Larry Sass SM ’94, PhD ’00 is an associate professor in the Department of Architecture. He earned his PhD and SM in architecture at MIT, and has a BArch from Pratt Institute in New York City. Sass joined the faculty in the Department of Architecture in 2002. His work focuses on the delivery of affordable housing for low-income families. He was included in an exhibit titled “Home Delivery: Fabricating the Modern Dwelling” at the Museum of Modern Art in New York City.

Sass’s teaching blends computation with design. His two signature courses, 4.500 (Design Computation: Art, Objects and Space) and 4.501 (Tiny Fab: Advancements in Rapid Design and Fabrication of Small Homes), reflect his specialization in digitally fabricating buildings and furniture from machines.

Professor and head of architecture Nicholas de Monchaux writes, “his classes provide crucial instruction and practice with 3D modeling and computer-generated rendering and animation …  [He] links digital design to fabrication, in a process that invites students to define desirable design attributes of an object, develop a digital model, prototype it, and construct it at full scale.”

More generally, Sass’ approach is to help students build confidence in their own design process through hands-on projects. MIT Class of 1942 Professor John Ochsendorf, MacVicar Fellow, and founding director of the Morningside Academy for Design with appointments in the departments of architecture and civil and environmental engineering, confirms, “Larry’s teaching is a perfect embodiment of the ‘mens et manus’ spirit … [he] requires his students to go back and forth from mind and hand throughout each design project.”

Students say that his classes are a journey of self-discovery, allowing them to learn more about themselves and their own abilities. Senior Natasha Hirt notes, “What I learned from Larry was not something one can glean from a textbook, but a new way of seeing space … he tectonically shifted my perspective on buildings. He also shifted my perspective on myself. I’m a better designer for his teachings, and perhaps more importantly, I better understand how I design.”

Senior Izzi Waitz echoes this sentiment: “Larry emphasizes the importance of intentionally thinking through your designs and being confident in your choices … he challenges, questions, and prompts you so that you learn to defend and support yourself on your own.”

As a UROP coordinator, Sass assures students that the “sky is the limit” and all ideas are welcome. Postgraduate teaching fellow and research associate Myles Sampson says, “During the last year of my SM program, I assisted Larry in conducting a year-long UROP project … He structured the learning experience in a way that allowed the students to freely flex their design muscles: no idea was too outrageous.”

Sass is equally devoted to his students outside the classroom. In his role as head of house at MacGregor House, he lives in community with more than 300 undergraduates each year, providing academic guidance, creating residential programs and recreational activities, and ensuring that student wellness and mental health is a No. 1 priority.

Professor of architecture and MacVicar Fellow Les Norford says, “In two significant ways, Larry has been ahead of his time: combining digital representation and design with making and being alert to the well-being of his students.”

“In his kindness, he honors the memory of Margaret MacVicar, as well as the spirit of MIT itself,” Hirt concludes. “He is a designer, a craftsman, and an innovator. He is an inspiration and a compass.”

On receiving this award, Sass is full of excitement: “I love teaching and being part of the MIT community. I am grateful for the opportunity to be part of the MacVicar family of fellows.”

New insights into training dynamics of deep classifiers

A new study from researchers at MIT and Brown University characterizes several properties that emerge during the training of deep classifiers, a type of artificial neural network commonly used for classification tasks such as image classification, speech recognition, and natural language processing.

The paper, “Dynamics in Deep Classifiers trained with the Square Loss: Normalization, Low Rank, Neural Collapse and Generalization Bounds,” published today in the journal Research, is the first of its kind to theoretically explore the dynamics of training deep classifiers with the square loss and how properties such as rank minimization, neural collapse, and dualities between the activation of neurons and the weights of the layers are intertwined.

In the study, the authors focused on two types of deep classifiers: fully connected deep networks and convolutional neural networks (CNNs).

A previous study examined the structural properties that develop in large neural networks at the final stages of training. That study focused on the last layer of the network and found that deep networks trained to fit a training dataset will eventually reach a state known as “neural collapse.” When neural collapse occurs, the network maps multiple examples of a particular class (such as images of cats) to a single template of that class. Ideally, the templates for each class should be as far apart from each other as possible, allowing the network to accurately classify new examples.

An MIT group based at the MIT Center for Brains, Minds and Machines studied the conditions under which networks can achieve neural collapse. Deep networks that have the three ingredients of stochastic gradient descent (SGD), weight decay regularization (WD), and weight normalization (WN) will display neural collapse if they are trained to fit their training data. The MIT group has taken a theoretical approach — as compared to the empirical approach of the earlier study — proving that neural collapse emerges from the minimization of the square loss using SGD, WD, and WN.

Co-author and MIT McGovern Institute postdoc Akshay Rangamani states, “Our analysis shows that neural collapse emerges from the minimization of the square loss with highly expressive deep neural networks. It also highlights the key roles played by weight decay regularization and stochastic gradient descent in driving solutions towards neural collapse.”

Weight decay is a regularization technique that prevents the network from over-fitting the training data by reducing the magnitude of the weights. Weight normalization scales the weight matrices of a network so that they have a similar scale. Low rank refers to a property of a matrix where it has a small number of non-zero singular values. Generalization bounds offer guarantees about the ability of a network to accurately predict new examples that it has not seen during training.

The authors found that the same theoretical observation that predicts a low-rank bias also predicts the existence of an intrinsic SGD noise in the weight matrices and in the output of the network. This noise is not generated by the randomness of the SGD algorithm but by an interesting dynamic trade-off between rank minimization and fitting of the data, which provides an intrinsic source of noise similar to what happens in dynamic systems in the chaotic regime. Such a random-like search may be beneficial for generalization because it may prevent over-fitting.

“Interestingly, this result validates the classical theory of generalization showing that traditional bounds are meaningful. It also provides a theoretical explanation for the superior performance in many tasks of sparse networks, such as CNNs, with respect to dense networks,” comments co-author and MIT McGovern Institute postdoc Tomer Galanti. In fact, the authors prove new norm-based generalization bounds for CNNs with localized kernels, that is a network with sparse connectivity in their weight matrices.

In this case, generalization can be orders of magnitude better than densely connected networks. This result validates the classical theory of generalization, showing that its bounds are meaningful, and goes against a number of recent papers expressing doubts about past approaches to generalization. It also provides a theoretical explanation for the superior performance of sparse networks, such as CNNs, with respect to dense networks. Thus far, the fact that CNNs and not dense networks represent the success story of deep networks has been almost completely ignored by machine learning theory. Instead, the theory presented here suggests that this is an important insight in why deep networks work as well as they do.

“This study provides one of the first theoretical analyses covering optimization, generalization, and approximation in deep networks and offers new insights into the properties that emerge during training,” says co-author Tomaso Poggio, the Eugene McDermott Professor at the Department of Brain and Cognitive Sciences at MIT and co-director of the Center for Brains, Minds and Machines. “Our results have the potential to advance our understanding of why deep learning works as well as it does.”

School of Science presents 2023 Infinite Expansion Awards

The MIT School of Science has announced seven postdocs and research scientists as recipients of the 2023 Infinite Expansion Award. Nominated by their peers and mentors, the awardees are recognized not only for their exceptional science, but for mentoring and advising junior colleagues, supporting educational programs, working with the MIT Postdoctoral Association, or contributing some other way to the Institute.

The 2023 Infinite Expansion award winners in the School of Science are:

  • Kyle Jenks, a postdoc in the Picower Institute for Learning and Memory, nominated by professor and Picower Institute investigator Mriganka Sur;
  • Matheus Victor, a postdoc in the Picower Institute, nominated by professor and Picower Institute director Li-Huei Tsai.

A monetary award is granted to recipients, and a celebratory reception will be held for the winners this spring with family, friends, nominators, and recipients of the Infinite Expansion Award.

Studies of unusual brains reveal critical insights into brain organization, function

EG (a pseudonym) is an accomplished woman in her early 60s: she is a college graduate and has an advanced professional degree. She has a stellar vocabulary—in the 98th percentile, according to tests—and has mastered a foreign language (Russian) to the point that she sometimes dreams in it.

She also has, likely since birth, been missing her left temporal lobe, a part of the brain known to be critical for language.

In 2016, EG contacted McGovern Institute Investigator Evelina Fedorenko, who studies the computations and brain regions that underlie language processing, to see if her team might be interested in including her in their research.

“EG didn’t know about her missing temporal lobe until age 25, when she had a brain scan for an unrelated reason,” says Fedorenko, the Frederick A. (1971) and Carole J. Middleton Career Development Associate Professor of Neuroscience at MIT. “As with many cases of early brain damage, she had no linguistic or cognitive deficits, but brains like hers are invaluable for understanding how cognitive functions reorganize in the tissue that remains.”

“I told her we definitely wanted to study her brain.” – Ev Fedorenko

Previous studies have shown that language processing relies on an interconnected network of frontal and temporal regions in the left hemisphere of the brain. EG’s unique brain presented an opportunity for Fedorenko’s team to explore how language develops in the absence of the temporal part of these core language regions.

Greta Tuckute, a graduate student in the Fedorenko lab, is the first author of the Neuropsychologia study. Photo: Caitlin Cunningham

Their results appeared recently in the journal Neuropsychologia. They found, for the first time, that temporal language regions appear to be critical for the emergence of frontal language regions in the same hemisphere — meaning, without a left temporal lobe, EG’s intact frontal lobe did not develop a capacity for language.

They also reveal much more: EG’s language system resides happily in her right hemisphere. “Our findings provide both visual and statistical proof of the brain’s remarkable plasticity, its ability to reorganize, in the face of extensive early damage,” says Greta Tuckute, a graduate student in the Fedorenko lab and first author of the paper.

In an introduction to the study, EG herself puts the social implications of the findings starkly. “Please do not call my brain abnormal, that creeps me out,” she . “My brain is atypical. If not for accidentally finding these differences, no one would pick me out of a crowd as likely to have these, or any other differences that make me unique.”

How we process language

The frontal and temporal lobes are part of the cerebrum, the largest part of the brain. The cerebrum controls many functions, including the five senses, language, working memory, personality, movement, learning, and reasoning. It is divided into two hemispheres, the left and the right, by a deep longitudinal fissure. The two hemispheres communicate via a thick bundle of nerve fibers called the corpus callosum. Each hemisphere comprises four main lobes—frontal, parietal, temporal, and occipital. Core parts of the language network reside in the frontal and temporal lobes.

Core parts of the language network (shown in teal) reside in the left frontal and temporal lobes. Image: Ev Fedorenko

In most individuals, the language system develops in both the right and left hemispheres, with the left side dominant from an early age. The frontal lobe develops slower than the temporal lobe. Together, the interconnected frontal and temporal language areas enable us to understand and produce words, phrases, and sentences.

How, then, did EG, with no left temporal lobe, come to speak, comprehend, and remember verbal information (even a foreign language!) with such proficiency?

Simply put, the right hemisphere took over: “EG has a completely well-functioning neurotypical-like language system in her right hemisphere,” says Tuckute. “It is incredible that a person can use a single hemisphere—and the right hemisphere at that, which in most people is not the dominant hemisphere where language is processed—and be perfectly fine.”

Journey into EG’s brain

In the study, the researchers conducted two scans of EG’s brain using functional magnetic resonance imaging (fMRI), one in 2016 and one in 2019, and had her complete a range of behaviorial tests. fMRI measures the level of blood oxygenation across the brain and can be used to make inferences about where neural activity is taking place. The researchers also scanned the brains of 151 “neurotypical” people. The large number of participants, combined with robust task paradigms and rigorous statistical analyses made it possible to draw conclusions from a single case such as EG.

Magnetic resonance image of EG’s brain showing missing left temporal lobe. Image: Fedorenko Lab

Fedorenko is a staunch advocate of the single case study approach—common in medicine but not currently in neuroscience. “Unusual brains—and unusual individuals more broadly—can provide critical insights into brain organization and function that we simply cannot gain by looking at more typical brains.” Studying individual brains with fMRI, however, requires paradigms that work robustly at the single-brain level. This is not true of most paradigms used in the field, which require averaging many brains together to obtain an effect. Developing individual-level fMRI paradigms for language research has been the focus of Fedorenko’s early work, although the main reason for doing so had nothing to do with studying atypical brains: individual-level analyses are simply better—they are more sensitive and their results are more interpretable and meaningful.

“Looking at high-quality data in an individual participant versus looking at a group-level map is akin to using a high-precision microscope versus looking with a naked myopic eye, when all you see is a blur,” she wrote in an article published in Current Opinion in Behaviorial Sciences in 2021. Having developed and validated such paradigms, though, is now allowing Fedorenko and her group to probe interesting brains.

While in the scanner, each participant performed a task that Fedorenko began developing more than a decade ago. They were presented with a series of words that form real, meaningful sentences, and with a series of “nonwords”—strings of letters that are pronounceable but without meaning. In typical brains, language areas respond more strongly when participants read sentences compared to when they read nonword sequences.

Similarly, in response to the real sentences, the language regions in EG’s right frontal and temporal lobes lit up—they were bursting with activity—while the left frontal lobe regions remained silent. In the neurotypical participants, the language regions in both the left and right frontal and temporal lobes lit up, with the left areas outshining the right.

fMRI showing EG’s language activation on the brain surface. The right frontal lobe shows robust activations, while the left frontal lobe does not have any language responsive areas. Image: Fedorenko lab

“EG showed a very strong response in the right temporal and frontal regions that process language,” says Tuckute. “And if you look at the controls, whose language dominant hemisphere is in the left, EG’s response in her right hemisphere was similar—or even higher—compared to theirs, just on the opposite side.”

Leaving no stone unturned, the researchers next asked whether the lack of language responses in EG’s left frontal lobe might be due to a general lack of response to cognitive tasks rather than just to language. So they conducted a non-language, working-memory task: they had EG and the neurotypical participants perform arithmetic addition problems while in the scanner. In typical brains, this task elicits responses in frontal and parietal areas in both hemisphers.

Not only did regions of EG’s right frontal lobe light up in response to the task, those in her left frontal lobe did, too. “Both EG’s language-dominant (right) hemisphere, and her non-language-dominant (left) hemisphere showed robust responses to this working-memory task ,” says Tuckute. “So, yes, there’s definitely cognitive processing going on there. This selective lack of language responses in EG’s left frontal lobe led us to conclude that, for language, you need the temporal language region to ‘wire up’ the frontal language region.”

Next steps

In science, the answer to one question opens the door to untold more. “In EG, language took over a large chunk of the right frontal and temporal lobes,” says Fedorenko. “So what happens to the functions that in neurotypical individuals generally live in the right hemisphere?”

Many of those, she says, are social functions. The team has already tested EG on social tasks and is currently exploring how those social functions cohabit with the language ones in her right hemisphere. How can they all fit? Do some of the social functions have to migrate to other parts of the brain? They are also working with EG’s family: they have now scanned EG’s three siblings (one of whom is missing most of her right temporal lobe; the other two are neurotypical) and her father (also neurotypical).

The “Interesting Brains Project” website details current projects, findings, and ways to participate.

The project has now grown to include many other individuals with interesting brains, who contacted Fedorenko after some of this work was covered by news outlets. A website for this project can be found here. The project promises to provide unique insights into how our plastic brains reorganize and adapt to various circumstances.

 

New collaboration aims to strengthen orthotic and prosthetic care in Sierra Leone

MIT’s K. Lisa Yang Center for Bionics has entered into a collaboration with the Government of Sierra Leone to strengthen the capabilities and services of that country’s orthotic and prosthetic (O&P) sector. Tens of thousands of people in Sierra Leone are in need of orthotic braces and artificial limbs, but access to such specialized medical care in this African nation is limited.

The agreement, reached between MIT, the Center for Bionics, and Sierra Leone’s Ministry of Health and Sanitation (MoHS), provides a detailed memorandum of understanding and intentions that will begin as a four-year program.  The collaborators aim to strengthen Sierra Leone’s O&P sector through six key objectives: data collection and clinic operations, education, supply chain, infrastructure, new technologies and mobile delivery of services.

Project Objectives

  1. Data Collection and Clinic Operations: collect comprehensive data on epidemiology, need, utilization, and access for O&P services across the country
  2. Education: create an inclusive education and training program for the people of Sierra Leone, to enable sustainable and independent operation of O&P services
  3. Supply Chain: establish supply chains for prosthetic and orthotic components, parts, and materials for fabrication of devices
  4. Infrastructure: prepare infrastructure (e.g., physical space, sufficient water, power and internet) to support increased production and services
  5. New Technologies: develop and translate innovative technologies with potential to improve O&P clinic operations and management, patient mobility, and the design or fabrication of devices
  6. Mobile Delivery: support outreach services and mobile delivery of care for patients in rural and difficult-to-reach areas

Working together, MIT’s bionics center and Sierra Leone’s MoHS aim to sustainably double the production and distribution of O&P services at Sierra Leone’s National Rehabilitation Centre and Bo Clinics over the next four years.

The team of MIT scientists who will be implementing this novel collaboration is led by Hugh Herr, MIT Professor of Media Arts and Sciences. Herr, himself a double amputee, serves as co-director of the K. Lisa Yang Center for Bionics, and heads the renowned Biomechatronics research group at the MIT Media Lab.

“From educational services, to supply chain, to new technology, this important MOU with the government of Sierra Leone will enable the Center to develop a broad, integrative approach to the orthotic and prosthetic sector within Sierra Leone, strengthening services and restoring much needed care to its citizens,” notes Professor Herr.

Sierra Leone’s Honorable Minister of Health Dr. Austin Demby also states: “As the Ministry of Health and Sanitation continues to galvanize efforts towards the attainment of Universal Health Coverage through the life stages approach, this collaboration will foster access, innovation and capacity building in the Orthotic and Prosthetic division. The ministry is pleased to work with and learn from MIT over the next four years in building resilient health systems, especially for vulnerable groups.”

“Our team at MIT brings together expertise across disciplines from global health systems to engineering and design,” added Francesca Riccio-Ackerman, the graduate student lead for the MIT Sierra Leone project. “This allows us to craft an innovative strategy with Sierra Leone’s Ministry of Health and Sanitation. Together we aim to improve available orthotic and prosthetic care for people with disabilities.”

The K. Lisa Yang Center for Bionics at the Massachusetts Institute of Technology pioneers transformational bionic interventions across a broad range of conditions affecting the body and mind. Based on fundamental scientific principles, the Center seeks to develop neural and mechanical interfaces for human-machine communications; integrate these interfaces into novel bionic platforms; perform clinical trials to accelerate the deployment of bionic products by the private sector; and leverage novel and durable, but affordable, materials and manufacturing processes to ensure equitable access to the latest bionic technology by all impacted individuals, especially those in developing countries. 

Sierra Leone’s Ministry of Health and Sanitation is responsible for health service delivery across the country, as well as regulation of the health sector to meet the health needs of its citizenry. 

For more information about this project, please visit: https://mitmedialab.info/prosforallproj2

 

How Huntington’s disease affects different neurons

In patients with Huntington’s disease, neurons in a part of the brain called the striatum are among the hardest-hit. Degeneration of these neurons contributes to patients’ loss of motor control, which is one of the major hallmarks of the disease.

Neuroscientists at MIT have now shown that two distinct cell populations in the striatum are affected differently by Huntington’s disease. They believe that neurodegeneration of one of these populations leads to motor impairments, while damage to the other population, located in structures called striosomes, may account for the mood disorders that are often see in the early stages of the disease.

“As many as 10 years ahead of the motor diagnosis, Huntington’s patients can experience mood disorders, and one possibility is that the striosomes might be involved in these,” says Ann Graybiel, an MIT Institute Professor, a member of MIT’s McGovern Institute for Brain Research, and one of the senior authors of the study.

Using single-cell RNA sequencing to analyze the genes expressed in mouse models of Huntington’s disease and postmortem brain samples from Huntington’s patients, the researchers found that cells of the striosomes and another structure, the matrix, begin to lose their distinguishing features as the disease progresses. The researchers hope that their mapping of the striatum and how it is affected by Huntington’s could help lead to new treatments that target specific cells within the brain.

This kind of analysis could also shed light on other brain disorders that affect the striatum, such as Parkinson’s disease and autism spectrum disorder, the researchers say.

Myriam Heiman, an associate professor in MIT’s Department of Brain and Cognitive Sciences and a member of the Picower Institute for Learning and Memory, and Manolis Kellis, a professor of computer science in MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) and a member of the Broad Institute of MIT and Harvard, are also senior authors of the study. Ayano Matsushima, a McGovern Institute research scientist, and Sergio Sebastian Pineda, an MIT graduate student, are the lead authors of the paper, which appears in Nature Communications.

Neuron vulnerability

Huntington’s disease leads to degeneration of brain structures called the basal ganglia, which are responsible for control of movement and also play roles in other behaviors, as well as emotions. For many years, Graybiel has been studying the striatum, a part of the basal ganglia that is involved in making decisions that require evaluating the outcomes of a particular action.

Many years ago, Graybiel discovered that the striatum is divided into striosomes, which are clusters of neurons, and the matrix, which surrounds the striosomes. She has also shown that striosomes are necessary for making decisions that require an anxiety-provoking cost-benefit analysis.

In a 2007 study, Richard Faull of the University of Auckland discovered that in postmortem brain tissue from Huntington’s patients, the striosomes showed a great deal of degeneration. Faull also found that while those patients were alive, many of them had shown signs of mood disorders such as depression before their motor symptoms developed.

To further explore the connections between the striatum and the mood and motor effects of Huntington’s, Graybiel teamed up with Kellis and Heiman to study the gene expression patterns of striosomal and matrix cells. To do that, the researchers used single-cell RNA sequencing to analyze human brain samples and brain tissue from two mouse models of Huntington’s disease.

Within the striatum, neurons can be classified as either D1 or D2 neurons. D1 neurons are involved in the “go” pathway, which initiates an action, and D2 neurons are part of the “no-go” pathway, which suppresses an action. D1 and D2 neurons can both be found within either the striosomes and the matrix.

The analysis of RNA expression in each of these types of cells revealed that striosomal neurons are harder hit by Huntington’s than matrix neurons. Furthermore, within the striosomes, D2 neurons are more vulnerable than D1.

The researchers also found that these four major cell types begin to lose their identifying molecular identities and become more difficult to distinguish from one another in Huntington’s disease. “Overall, the distinction between striosomes and matrix becomes really blurry,” Graybiel says.

Striosomal disorders

The findings suggest that damage to the striosomes, which are known to be involved in regulating mood, may be responsible for the mood disorders that strike Huntington’s patients in the early stages of the disease. Later on, degeneration of the matrix neurons likely contributes to the decline of motor function, the researchers say.

In future work, the researchers hope to explore how degeneration or abnormal gene expression in the striosomes may contribute to other brain disorders.

Previous research has shown that overactivity of striosomes can lead to the development of repetitive behaviors such as those seen in autism, obsessive compulsive disorder, and Tourette’s syndrome. In this study, at least one of the genes that the researchers discovered was overexpressed in the striosomes of Huntington’s brains is also linked to autism.

Additionally, many striosome neurons project to the part of the brain that is most affected by Parkinson’s disease (the substantia nigra, which produces most of the brain’s dopamine).

“There are many, many disorders that probably involve the striatum, and now, partly through transcriptomics, we’re working to understand how all of this could fit together,” Graybiel says.

The research was funded by the Saks Kavanaugh Foundation, the CHDI Foundation, the National Institutes of Health, the Nancy Lurie Marks Family Foundation, the Simons Foundation, the JPB Foundation, the Kristin R. Pressman and Jessica J. Pourian ’13 Fund, and Robert Buxton.

Self-assembling proteins can store cellular “memories”

As cells perform their everyday functions, they turn on a variety of genes and cellular pathways. MIT engineers have now coaxed cells to inscribe the history of these events in a long protein chain that can be imaged using a light microscope.

Cells programmed to produce these chains continuously add building blocks that encode particular cellular events. Later, the ordered protein chains can be labeled with fluorescent molecules and read under a microscope, allowing researchers to reconstruct the timing of the events.

This technique could help shed light on the steps that underlie processes such as memory formation, response to drug treatment, and gene expression.

“There are a lot of changes that happen at organ or body scale, over hours to weeks, which cannot be tracked over time,” says Edward Boyden, the Y. Eva Tan Professor in Neurotechnology, a professor of biological engineering and brain and cognitive sciences at MIT, a Howard Hughes Medical Institute investigator, and a member of MIT’s McGovern Institute for Brain Research and Koch Institute for Integrative Cancer Research.

If the technique could be extended to work over longer time periods, it could also be used to study processes such as aging and disease progression, the researchers say.

Boyden is the senior author of the study, which appears today in Nature Biotechnology. Changyang Linghu, a former J. Douglas Tan Postdoctoral Fellow at the McGovern Institute, who is now an assistant professor at the University of Michigan, is the lead author of the paper.

Cellular history

Biological systems such as organs contain many different kinds of cells, all of which have distinctive functions. One way to study these functions is to image proteins, RNA, or other molecules inside the cells, which provide hints to what the cells are doing. However, most methods for doing this offer only a glimpse of a single moment in time, or don’t work well with very large populations of cells.

“Biological systems are often composed of a large number of different types of cells. For example, the human brain has 86 billion cells,” Linghu says. “To understand those kinds of biological systems, we need to observe physiological events over time in these large cell populations.”

To achieve that, the research team came up with the idea of recording cellular events as a series of protein subunits that are continuously added to a chain. To create their chains, the researchers used engineered protein subunits, not normally found in living cells, that can self-assemble into long filaments.

The researchers designed a genetically encoded system in which one of these subunits is continuously produced inside cells, while the other is generated only when a specific event occurs. Each subunit also contains a very short peptide called an epitope tag — in this case, the researchers chose tags called HA and V5. Each of these tags can bind to a different fluorescent antibody, making it easy to visualize the tags later on and determine the sequence of the protein subunits.

For this study, the researchers made production of the V5-containing subunit contingent on the activation of a gene called c-fos, which is involved in encoding new memories. HA-tagged subunits make up most of the chain, but whenever the V5 tag shows up in the chain, that means that c-fos was activated during that time.

“We’re hoping to use this kind of protein self-assembly to record activity in every single cell,” Linghu says. “It’s not only a snapshot in time, but also records past history, just like how tree rings can permanently store information over time as the wood grows.”

Recording events

In this study, the researchers first used their system to record activation of c-fos in neurons growing in a lab dish. The c-fos gene was activated by chemically induced activation of the neurons, which caused the V5 subunit to be added to the protein chain.

To explore whether this approach could work in the brains of animals, the researchers programmed brain cells of mice to generate protein chains that would reveal when the animals were exposed to a particular drug. Later, the researchers were able to detect that exposure by preserving the tissue and analyzing it with a light microscope.

The researchers designed their system to be modular, so that different epitope tags can be swapped in, or different types of cellular events can be detected, including, in principle, cell division or activation of enzymes called protein kinases, which help control many cellular pathways.

The researchers also hope to extend the recording period that they can achieve. In this study, they recorded events for several days before imaging the tissue. There is a tradeoff between the amount of time that can be recorded and the time resolution, or frequency of event recording, because the length of the protein chain is limited by the size of the cell.

“The total amount of information it could store is fixed, but we could in principle slow down or increase the speed of the growth of the chain,” Linghu says. “If we want to record for a longer time, we could slow down the synthesis so that it will reach the size of the cell within, let’s say two weeks. In that way we could record longer, but with less time resolution.”

The researchers are also working on engineering the system so that it can record multiple types of events in the same chain, by increasing the number of different subunits that can be incorporated.

The research was funded by the Hock E. Tan and K. Lisa Yang Center for Autism Research, John Doerr, the National Institutes of Health, the National Science Foundation, the U.S. Army Research Office, and the Howard Hughes Medical Institute.