Regina Barzilay, School of Engineering Distinguished Professor for AI and Health at MIT, CSAIL Principal Investigator, and Jameel Clinic AI Faculty Lead, has been awarded the 2025 Frances E. Allen Medal from the Institute of Electrical and Electronics Engineers (IEEE). Barzilay’s award recognizes the impact of her machine-learning algorithms on medicine and natural language processing.
Whether you’re describing the sound of your faulty car engine or meowing like your neighbor’s cat, imitating sounds with your voice can be a helpful way to relay a concept when words don’t do the trick.
Generative AI systems like large language models rely heavily on deep learning - and, in particular, transformers. Transformers make use of an “attention mechanism” for modeling interactions among inputs, which essentially involves doing nonlinear pairwise comparison between inputs and assigning different weights to tokens in a sequence, enabling a prioritization of some over others. The empirical effectiveness of this attention mechanism has led some in the community to claim that attention is “all you need” (the title of the original 2017 Google paper that introduced transformers).