We are developing tools to enable physically-embodied, spatial AI as a "technological superpower" for Robots and Humans. Semantic simultaneous localization and mapping can enable long-lived autonomous systems to navigate in complex dynamic environments. Our dream is to create mobile robots that can build and maintain models of the world through lifelong learning, improving their performance over time, and helping humans to perform difficult tasks.
Large language models (LLMs) are becoming increasingly useful for programming and robotics tasks, but for more complicated reasoning problems, the gap between these systems and humans looms large. Without the ability to learn new concepts like humans do, these systems fail to form good abstractions — essentially, high-level representations of complex concepts that skip less-important details — and thus sputter when asked to do more sophisticated tasks.
For nearly a decade, a team of MIT Computer Science and Artificial Intelligence Laboratory (CSAIL) researchers have been seeking to uncover why certain images persist in a people's minds, while many others fade. To do this, they set out to map the spatio-temporal brain dynamics involved in recognizing a visual image. And now for the first time, scientists harnessed the combined strengths of magnetoencephalography (MEG), which captures the timing of brain activity, and functional magnetic resonance imaging (fMRI), which identifies active brain regions, to precisely determine when and where the brain processes a memorable image.
To build AI systems that can collaborate effectively with humans, it helps to have a good model of human behavior to start with. But humans tend to behave suboptimally when making decisions.
Daniela Rus is a pioneering roboticist and a professor of electrical engineering and computer science at MIT. She is the director of the Computer Science and Artificial Intelligence Laboratory. She is also a member of the National Academy of Engineering, the American Academy of Arts and Sciences, and a MacArthur Fellow.
In biomedicine, segmentation involves annotating pixels from an important structure in a medical image, like an organ or cell. Artificial intelligence models can help clinicians by highlighting pixels that may show signs of a certain disease or anomaly.
A user could ask ChatGPT to write a computer program or summarize an article, and the AI chatbot would likely be able to generate useful code or write a cogent synopsis. However, someone could also ask for instructions to build a bomb, and the chatbot might be able to provide those, too.