Image
alt="The automated, multimodal approach developed by MIT researchers interprets artificial vision models that evaluate the properties of images (Credits: iStock)."
CSAIL article

As artificial intelligence models become increasingly prevalent and are integrated into diverse sectors like health care, finance, education, transportation, and entertainment, understanding how they work under the hood is critical. Interpreting the mechanisms underlying AI models enables us to audit them for safety and biases, with the potential to deepen our understanding of the science behind intelligence itself.

Image
MosaicML (L-R): Naveen Rao, Michael Carbin, Julie Shin Choi, Jonathan Frankle, and Hanlin Tang (Credit: Courtesy of MosaicML).
CSAIL article

The impact of artificial intelligence will never be equitable if there’s only one company that builds and controls the models (not to mention the data that go into them). Unfortunately, today’s AI models are made up of billions of parameters that must be trained and tuned to maximize performance for each use case, putting the most powerful AI models out of reach for most people and companies.

Image
Ryan Williams headshot
CSAIL article

The European Association for Theoretical Computer Science (EATCS) recently awarded Ryan Williams, MIT EECS professor and CSAIL member, with the 2024 Gödel Prize for his 2011 paper, “Non-Uniform ACC Circuit Lower Bounds.” Williams receives this honor for presenting a novel paradigm for a “rich two-way connection" between algorithmic techniques and lower-bound methods.

Image
To close the gap with classical computers, researchers created the quantum control machine — an instruction set for a quantum computer that works like the classical idea of a virtual machine (Credits: Alex Shipps/MIT CSAIL).
CSAIL article

When MIT professor and now Computer Science and Artificial Intelligence Laboratory (CSAIL) member Peter Shor first demonstrated the potential of quantum computers to solve problems faster than classical ones, he inspired scientists to imagine countless possibilities for the emerging technology. Thirty years later, though, the quantum edge remains a peak not yet reached.

Image
Researchers from MIT and elsewhere found that complex large language machine-learning models use a simple mechanism to retrieve stored knowledge when they respond to a user prompt. The researchers can leverage these simple mechanisms to see what the model knows about different subjects, and also possibly correct false information that it has stored (Credits: iStock).
CSAIL article

Large language models, such as those that power popular artificial intelligence chatbots like ChatGPT, are incredibly complex. Even though these models are being used as tools in many areas, such as customer support, code generation, and language translation, scientists still don’t fully grasp how they work.