Daniela Rus, Director of CSAIL and MIT EECS Professor, recently received the 2025 Edison Medal from the Institute of Electrical and Electronics Engineers (IEEE). The award recognizes her leadership and pioneering work in modern robotics.
Daniela Rus, Director of CSAIL and MIT EECS Professor, was recently named a co-recipient of the 2024 John Scott Award by the Board of Directors of City Trusts. This prestigious honor, steeped in historical significance, celebrates scientific innovation at the very location where American independence was signed in Philadelphia, a testament to the enduring connection between scientific progress and human potential.
For roboticists, one challenge towers above all others: generalization – the ability to create machines that can adapt to any environment or condition. Since the 1970s, the field has evolved from writing sophisticated programs to using deep learning, teaching robots to learn directly from human behavior. But a critical bottleneck remains: data quality. To improve, robots need to encounter scenarios that push the boundaries of their capabilities, operating at the edge of their mastery.
In the classic cartoon “The Jetsons,” Rosie the robotic maid seamlessly switches from vacuuming the house to cooking dinner to taking out the trash. But in real life, training a general-purpose robot remains a major challenge.
In the current AI zeitgeist, sequence models have skyrocketed in popularity for their ability to analyze data and predict what to do next. For instance, you’ve likely used next-token prediction models like ChatGPT, which anticipate each word (token) in a sequence to form answers to users’ queries. There are also full-sequence diffusion models like Sora, which convert words into dazzling, realistic visuals by successively “denoising” an entire video sequence
The Proceedings of the ACM on Interactive, Mobile, Wearable, and Ubiquitous Technologies (IMWUT) Editorial Board has awarded MIT Computer Science and Artificial Intelligence Laboratory (CSAIL) and the Gwangju Institute of Science and Technology (GIST) researchers with a Distinguished Paper Award for their evaluation of visual explanations in autonomous vehicles’ decision-making.
For robots, simulation is a great teacher for learning long-horizon (multi-step) tasks — especially compared to how long it takes to collect real-world training data.
Imagine you’re tasked with sending a team of football players onto a field to assess the condition of the grass (a likely task for them, of course). If you pick their positions randomly, they might cluster together in some areas while completely neglecting others. But if you give them a strategy, like spreading out uniformly across the field, you might get a far more accurate picture of the grass condition.