Papers Code Mit Ibm Watson Ai Lab
Peer-review is the lifeblood of scientific validation and a guardrail against runaway hype in AI. Our commitment to publishing in the top venues reflects our grounding in what is real, reproducible, and truly innovative. CLEVRER: The first video dataset for neuro-symbolic reasoning Computer Vision Explainability ICLR Neuro-Symbolic AI We are a community of scientists at MIT and IBM Research. We conduct AI research and work with global organizations to bridge algorithms to impact business and society.
Dina Katabi elected to the National Academy of Medicine Method teaches generative AI models to locate personalized objects Most languages use word position and sentence structure to extract meaning. For example, “The cat sat on the box,” is not the same as “The box was on the cat.” Over a long text, like a financial document or a novel, the syntax of these... Similarly, a person might be tracking variables in a piece of code or following instructions that have conditional actions. These are examples of state changes and sequential reasoning that we expect state-of-the-art artificial intelligence systems to excel at; however, the existing, cutting-edge attention mechanism within transformers — the primarily architecture used in large...
An attention mechanism allows an LLM to look back at earlier parts of a query or document and, based on its training, determine which details and words matter most; however, this mechanism alone does... It “sees” all of the input words, a.k.a. tokens, at the same time and handles them in the order that they’re presented, so researchers have developed techniques to encode position information. This is key for domains that are highly structured, like language. But the predominant position-encoding method, called rotary position encoding (RoPE), only takes into account the relative distance between tokens in a sequence and is independent of the input data. This means that, for example, words that are four positions apart, like “cat” and “box” in the example above, will all receive the same fixed mathematical rotation specific to that relative distance.
Now research led by MIT and the MIT-IBM Watson AI Lab has produced an encoding technique known as “PaTH Attention” that makes positional information adaptive and context-aware rather than static, as with RoPE. “Transformers enable accurate and scalable modeling of many domains, but they have these limitations vis-a-vis state tracking, a class of phenomena that is thought to underlie important capabilities that we want in our AI... So, the important question is: How can we maintain the scalability and efficiency of transformers, while enabling state tracking?” says the paper’s senior author Yoon Kim, an associate professor in the Department of Electrical... A new way to increase the capabilities of large language models Nuno Loureiro, professor and director of MIT’s Plasma Science and Fusion Center, dies at 47 Enabling small language models to solve complex reasoning tasks
MIT affiliates named 2025 Schmidt Sciences AI2050 Fellows A smarter way for large language models to think about hard problems AI is a big topic. We find it helpful to modify it with some adjectives: Narrow, Broad, General. Narrow AI is the ability to perform specific tasks at a super-human rate within various categories, from chess, Jeopardy!, and Go, to voice assistance, debate, language translation, and image classification. Broad AI is next.
We’re just entering this frontier, but when it’s fully realized, it will feature AI systems that use and integrate multimodal data streams, learn more efficiently and flexibly, and traverse multiple tasks and domains. Broad AI will have powerful implications for business and society. Finally, General AI is essentially what science fiction has long imagined: AI systems capable of complex reasoning and full autonomy. Some scientists estimate that General AI could be possible sometime around 2050 – which is really little more than guesswork. Others say it will never be possible. For now, we’re focused on leading the next generation of Broad AI technologies for the betterment of business and society.
Here are some of the key technical themes shaping the path to Broad AI. CSAIL researchers find even “untrainable” neural nets can learn effectively when guided by another network’s built-in biases using their guidance method. MIT-IBM Watson AI Lab researchers developed an expressive architecture that provides better state tracking and sequential reasoning in LLMs over long texts. The “self-steering” DisCIPL system directs small models to work together on tasks with constraints, like itinerary planning and budgeting. This new technique enables LLMs to dynamically adjust the amount of computation they use for reasoning, based on the difficulty of the question. MIT PhD students who interned with the MIT-IBM Watson AI Lab Summer Program are pushing AI tools to be more flexible, efficient, and grounded in truth.
Building Efficient AI Ask the Experts February 22, 2024 | 12-12:30 PM ET Join MIT Associate Professor and Lab researcher Jonathan Ragan-Kelley in a virtual discussion moderated by Aude Oliva, Lab Co-Director. In order to perform a particular task and “work smarter, not harder,” the right tool needs to be used; the same goes for AI models. This means searching for programs that can solve a task and training them by optimizing their parameters, through multi-objective optimization of models that have trade-offs between task performance/accuracy and computational cost, particularly by designing... These programs can generate a set of efficient solutions for the models and pair that with the appropriate hardware to make deep learning more cost-effective and memory efficient to run faster and on smaller... Registration is required.
Speaker Jonathan Ragan-Kelley, PhD is an associate professor in MIT’s Department of Electrical Engineering and Computer Science, and a principal investigator at the Computer Science and Artificial Intelligence Laboratory. His work focuses on high-efficiency computer graphics, where graphics intersect with systems, architecture, and compilers. Before coming to MIT, Ragan-Kelley was an assistant professor of computer science at the University of California, Berkeley, a postdoc at Stanford, and a visiting researcher at Google. He has worked at three major GPU vendors in architecture, compilers, and research, and built a real-time preview system for the special effects industry in collaboration with Industrial Light & Magic. He earned a PhD from MIT, where he helped develop the Halide programming language for image-processing.
People Also Search
- Papers + Code - Code - MIT-IBM Watson AI Lab
- Home - MIT-IBM Watson AI Lab
- Search - MIT-IBM Watson AI Lab
- A new way to increase the capabilities of large language models | MIT ...
- News - MIT-IBM Watson AI Lab
- MIT-IBM Watson AI Lab - Home - ACM Digital Library
- PDF Report to the President year ended June 30, 2025, MIT-IBM Watson AI Lab
- Research - MIT-IBM Watson AI Lab
- MIT-IBM Watson AI Lab | MIT News | Massachusetts Institute of Technology
- Building Efficient AI - MIT-IBM Watson AI Lab
Peer-review Is The Lifeblood Of Scientific Validation And A Guardrail
Peer-review is the lifeblood of scientific validation and a guardrail against runaway hype in AI. Our commitment to publishing in the top venues reflects our grounding in what is real, reproducible, and truly innovative. CLEVRER: The first video dataset for neuro-symbolic reasoning Computer Vision Explainability ICLR Neuro-Symbolic AI We are a community of scientists at MIT and IBM Research. We co...
Dina Katabi Elected To The National Academy Of Medicine Method
Dina Katabi elected to the National Academy of Medicine Method teaches generative AI models to locate personalized objects Most languages use word position and sentence structure to extract meaning. For example, “The cat sat on the box,” is not the same as “The box was on the cat.” Over a long text, like a financial document or a novel, the syntax of these... Similarly, a person might be tracking ...
An Attention Mechanism Allows An LLM To Look Back At
An attention mechanism allows an LLM to look back at earlier parts of a query or document and, based on its training, determine which details and words matter most; however, this mechanism alone does... It “sees” all of the input words, a.k.a. tokens, at the same time and handles them in the order that they’re presented, so researchers have developed techniques to encode position information. This...
Now Research Led By MIT And The MIT-IBM Watson AI
Now research led by MIT and the MIT-IBM Watson AI Lab has produced an encoding technique known as “PaTH Attention” that makes positional information adaptive and context-aware rather than static, as with RoPE. “Transformers enable accurate and scalable modeling of many domains, but they have these limitations vis-a-vis state tracking, a class of phenomena that is thought to underlie important capa...
MIT Affiliates Named 2025 Schmidt Sciences AI2050 Fellows A Smarter
MIT affiliates named 2025 Schmidt Sciences AI2050 Fellows A smarter way for large language models to think about hard problems AI is a big topic. We find it helpful to modify it with some adjectives: Narrow, Broad, General. Narrow AI is the ability to perform specific tasks at a super-human rate within various categories, from chess, Jeopardy!, and Go, to voice assistance, debate, language transla...