As part of our multi-year collaboration with Liverpool FC, we develop a full AI system that can advise coaches on corner kicks
As part of our multi-year collaboration with Liverpool FC, we develop a full AI system that can advise coaches on corner kicks
As part of our multi-year collaboration with Liverpool FC, we develop a full AI system that can advise coaches on corner kicks
Introducing SIMA, a Scalable Instructable Multiworld Agent
Introducing SIMA, a Scalable Instructable Multiworld Agent
Introducing SIMA, a Scalable Instructable Multiworld Agent
Introducing SIMA, a Scalable Instructable Multiworld Agent
Introducing SIMA, a Scalable Instructable Multiworld Agent
Introducing SIMA, a Scalable Instructable Multiworld Agent
Every year, the Berkeley Artificial Intelligence Research (BAIR) Lab graduates some of the most talented and innovative minds in artificial intelligence and machine learning. Our Ph.D. graduates have each expanded the frontiers of AI research and are now ready to embark on new adventures in academia, industry, and beyond.
These fantastic individuals bring with them a wealth of knowledge, fresh ideas, and a drive to continue contributing to the advancement of AI. Their work at BAIR, ranging from deep learning, robotics, and natural language processing to computer vision, security, and much more, has contributed significantly to their fields and has had transformative impacts on society.
This website is dedicated to showcasing our colleagues, making it easier for academic institutions, research organizations, and industry leaders to discover and recruit from the newest generation of AI pioneers. Here, you’ll find detailed profiles, research interests, and contact information for each of our graduates. We invite you to explore the potential collaborations and opportunities these graduates present as they seek to apply their expertise and insights in new environments.
Join us in celebrating the achievements of BAIR’s latest PhD graduates. Their journey is just beginning, and the future they will help build is bright!
Thank you to our friends at the Stanford AI Lab for this idea!
Abdus Salam Azad
Email: salam_azad@berkeley.edu
Website: https://www.azadsalam.org/
Advisor(s): Ion Stoica
Research Blurb: My research interest lies broadly in the field of Machine Learning and Artificial Intelligence. During my PhD I have focused on Environment Generation/ Curriculum Learning methods for training Autonomous Agents with Reinforcement Learning. Specifically, I work on methods that algorithmically generates diverse training environments (i.e., learning scenarios) for autonomous agents to improve generalization and sample efficiency. Currently, I am working on Large Language Model (LLM) based autonomous agents.
Jobs Interested In: Research Scientist, ML Engineer
Alicia Tsai
Email: aliciatsai@berkeley.edu
Website: https://www.aliciatsai.com/
Advisor(s): Laurent El Ghaoui
Research Blurb: My research delves into the theoretical aspects of deep implicit models, beginning with a unified "state-space" representation that simplifies notation. Additionally, my work explores various training challenges associated with deep learning, including problems amenable to convex and non-convex optimization. In addition to theoretical exploration, my research extends the potential applications to various problem domains, including natural language processing, and natural science.
Jobs Interested In: Research Scientist, Applied Scientist, Machine Learning Engineer
Catherine Weaver
Email: catherine22@berkeley.edu
Website: https://cwj22.github.io
Advisor(s): Masayoshi Tomizuka, Wei Zhan
Research Blurb: My research focuses on machine learning and control algorithms for the challenging task of autonomous racing in Gran Turismo Sport. I leverage my background in Mechanical Engineering to discover how machine learning and model-based optimal control can create safe, high-performance control systems for robotics and autonomous systems. A particular emphasis of mine has been how to leverage offline datasets (e.g. human player's racing trajectories) to inform better, more sample efficient control algorithms.
Jobs Interested In: Research Scientist and Robotics/Controls Engineer
Chawin Sitawarin
Email: chawin.sitawarin@gmail.com
Website: https://chawins.github.io/
Advisor(s): David Wagner
Research Blurb: I am broadly interested in the security and safety aspects of machine learning systems. Most of my previous works are in the domain of adversarial machine learning, particularly adversarial examples and robustness of machine learning algorithms. More recently, I am excited about emerging security and privacy risks on large language models.
Jobs Interested In: Research scientist
Dhruv Shah
Email: shah@cs.berkeley.edu
Website: http://cs.berkeley.edu/~shah/
Advisor(s): Sergey Levine
Research Blurb: I train big(-ish) models and make robots smarter.
Jobs Interested In: Research scientist, roboticist
Eliza Kosoy
Email: eko@berkeley.edu
Website: https://www.elizakosoy.com/
Advisor(s): Alison Gopnik
Research Blurb: Eliza Kosoy works at the intersection of child development and AI with Prof. Alison Gopnik. Her work includes creating evaluative benchmarks for LLMs rooted in child development and studying how children and adults use GenAI models such as ChatGPT/Dalle and form mental models about them. She’s an intern at Google working on the AI/UX team and previously with the Empathy Lab. She has published in Neurips, ICML, ICLR, Cogsci and cognition. Her thesis work created a unified virtual environment for testing children and AI models in one place for the purposes of training RL models. She also has experience building startups and STEM hardware coding toys.
Jobs Interested In: Research Scientist (child development and AI), AI safety (specializing in children), User Experience (UX) Researcher (specializing in mixed methods, youth, AI, LLMs), Education and AI (STEM toys)
Fangyu Wu
Email: fangyuwu@berkeley.edu
Website: https://fangyuwu.com/
Advisor(s): Alexandre Bayen
Research Blurb: Under the mentorship of Prof. Alexandre Bayen, Fangyu focuses on the application of optimization methods to multi-agent robotic systems, particularly in the planning and control of automated vehicles.
Jobs Interested In: Faculty, or research scientist in control, optimization, and robotics
Frances Ding
Email: frances@berkeley.edu
Website: https://www.francesding.com/
Advisor(s): Jacob Steinhardt, Moritz Hardt
Research Blurb: My research focus is in machine learning for protein modeling. I work on improving protein property classification and protein design, as well as understanding what different protein models learn. I have previously worked on sequence models for DNA and RNA, and benchmarks for evaluating the interpretability and fairness of ML models across domains.
Jobs Interested In: Research scientist
Jianlan Luo
Email: jianlanluo@eecs.berkeley.edu
Website: https://people.eecs.berkeley.edu/~jianlanluo/
Advisor(s): Sergey Levine
Research Blurb: My research interests are broadly in scalable algorithms and practice of machine learning, robotics, and controls; particularly their intersections.
Jobs Interested In: Faculty, Research Scientist
Kathy Jang
Email: kathyjang@gmail.com
Website: https://kathyjang.com
Advisor(s): Alexandre Bayen
Research Blurb: My thesis work has specialized in reinforcement learning for autonomous vehicles, focusing on enhancing decision-making and efficiency in applied settings. In future work, I'm eager to apply these principles to broader challenges across domains like natural language processing. With my background, my aim is to see the direct impact of my efforts by contributing to innovative AI research and solutions.
Jobs Interested In: ML research scientist/engineer
Kevin Lin
Email: k-lin@berkeley.edu
Website: https://people.eecs.berkeley.edu/~kevinlin/
Advisor(s): Dan Klein, Joseph E. Gonzalez
Research Blurb: My research focuses on understanding and improving how language models use and provide information.
Jobs Interested In: Research Scientist
Nikhil Ghosh
Email: nikhil_ghosh@berkeley.edu
Website: https://nikhil-ghosh-berkeley.github.io/
Advisor(s): Bin Yu, Song Mei
Research Blurb: I am interested in developing a better foundational understanding of deep learning and improving practical systems, using both theoretical and empirical methodology. Currently, I am especially interested in improving the efficiency of large models by studying how to properly scale hyperparameters with model size.
Jobs Interested In: Research Scientist
Olivia Watkins
Email: oliviawatkins@berkeley.edu
Website: https://aliengirlliv.github.io/oliviawatkins
Advisor(s): Pieter Abbeel and Trevor Darrell
Research Blurb: My work involves RL, BC, learning from humans, and using common-sense foundation model reasoning for agent learning. I’m excited about language agent learning, supervision, alignment & robustness.
Jobs Interested In: Research scientist
Ruiming Cao
Email: rcao@berkeley.edu
Website: https://rmcao.net
Advisor(s): Laura Waller
Research Blurb: My research is on computational imaging, particularly the space-time modeling for dynamic scene recovery and motion estimation. I also work on optical microscopy techniques, optimization-based optical design, event camera processing, novel view rendering.
Jobs Interested In: Research scientist, postdoc, faculty
Ryan Hoque
Email: ryanhoque@berkeley.edu
Website: https://ryanhoque.github.io
Advisor(s): Ken Goldberg
Research Blurb: Imitation learning and reinforcement learning algorithms that scale to large robot fleets performing manipulation and other complex tasks.
Jobs Interested In: Research Scientist
Sam Toyer
Email: sdt@berkeley.edu
Website: https://www.qxcv.net/
Advisor(s): Stuart Russell
Research Blurb: My research focuses on making language models secure, robust and safe. I also have experience in vision, planning, imitation learning, reinforcement learning, and reward learning.
Jobs Interested In: Research scientist
Shishir G. Patil
Email: shishirpatil2007@gmail.com
Website: https://shishirpatil.github.io/
Advisor(s): Joseph Gonzalez
Research Blurb: Gorilla LLM - Teaching LLMs to use tools (https://gorilla.cs.berkeley.edu/); LLM Execution Engine: Guaranteeing reversibility, robustness, and minimizing blast-radius for LLM-Agents incorporated into user and enterprise workflows; POET: Memory bound, and energy efficient fine-tuning of LLMs on edge devices such as smartphones and laptops (https://poet.cs.berkeley.edu/).
Jobs Interested In: Research Scientist
Suzie Petryk
Email: spetryk@berkeley.edu
Website: https://suziepetryk.com/
Advisor(s): Trevor Darrell, Joseph Gonzalez
Research Blurb: I work on improving the reliability and safety of multimodal models. My focus has been on localizing and reducing hallucinations for vision + language models, along with measuring and using uncertainty and mitigating bias. My interests lay in applying solutions to these challenges in actual production scenarios, rather than solely in academic environments.
Jobs Interested In: Applied research scientist in generative AI, safety, and/or accessibility
Xingyu Lin
Email: xingyu@berkeley.edu
Website: https://xingyu-lin.github.io/
Advisor(s): Pieter Abbeel
Research Blurb: My research lies in robotics, machine learning, and computer vision, with the primary goal of learning generalizable robot skills from two angles: (1) Learning structured world models with spatial and temporal abstractions. (2) Pre-training visual representation and skills to enable knowledge transfer from Internet-scale vision datasets and simulators.
Jobs Interested In: Faculty, or research scientist
Yaodong Yu
Email: yyu@eecs.berkeley.edu
Website: https://yaodongyu.github.io/
Advisor(s): Michael I. Jordan, Yi Ma
Research Blurb: My research interests are broadly in theory and practice of trustworthy machine learning, including interpretability, privacy, and robustness.
Jobs Interested In: Faculty
Gemma is built for responsible AI development from the same research and technology used to create Gemini models.
Gemma is built for responsible AI development from the same research and technology used to create Gemini models.
Gemma is built for responsible AI development from the same research and technology used to create Gemini models.
Gemma is built for responsible AI development from the same research and technology used to create Gemini models.
Gemma is built for responsible AI development from the same research and technology used to create Gemini models.
Gemma is built for responsible AI development from the same research and technology used to create Gemini models.
AI caught everyone’s attention in 2023 with Large Language Models (LLMs) that can be instructed to perform general tasks, such as translation or coding, just by prompting. This naturally led to an intense focus on models as the primary ingredient in AI application development, with everyone wondering what capabilities new LLMs will bring.
As more developers begin to build using LLMs, however, we believe that this focus is rapidly changing: state-of-the-art AI results are increasingly obtained by compound systems with multiple components, not just monolithic models.
For example, Google’s AlphaCode 2 set state-of-the-art results in programming through a carefully engineered system that uses LLMs to generate up to 1 million possible solutions for a task and then filter down the set. AlphaGeometry, likewise, combines an LLM with a traditional symbolic solver to tackle olympiad problems. In enterprises, our colleagues at Databricks found that 60% of LLM applications use some form of retrieval-augmented generation (RAG), and 30% use multi-step chains.
Even researchers working on traditional language model tasks, who used to report results from a single LLM call, are now reporting results from increasingly complex inference strategies: Microsoft wrote about a chaining strategy that exceeded GPT-4’s accuracy on medical exams by 9%, and Google’s Gemini launch post measured its MMLU benchmark results using a new CoT@32 inference strategy that calls the model 32 times, which raised questions about its comparison to just a single call to GPT-4. This shift to compound systems opens many interesting design questions, but it is also exciting, because it means leading AI results can be achieved through clever engineering, not just scaling up training.
In this post, we analyze the trend toward compound AI systems and what it means for AI developers. Why are developers building compound systems? Is this paradigm here to stay as models improve? And what are the emerging tools for developing and optimizing such systems—an area that has received far less research than model training? We argue that compound AI systems will likely be the best way to maximize AI results in the future, and might be one of the most impactful trends in AI in 2024.
Increasingly many new AI results are from compound systems.
Why Use Compound AI Systems?
We define a Compound AI System as a system that tackles AI tasks using multiple interacting components, including multiple calls to models, retrievers, or external tools. In contrast, an AI Model is simply a statistical model, e.g., a Transformer that predicts the next token in text.
Even though AI models are continually getting better, and there is no clear end in sight to their scaling, more and more state-of-the-art results are obtained using compound systems. Why is that? We have seen several distinct reasons:
Some tasks are easier to improve via system design. While LLMs appear to follow remarkable scaling laws that predictably yield better results with more compute, in many applications, scaling offers lower returns-vs-cost than building a compound system. For example, suppose that the current best LLM can solve coding contest problems 30% of the time, and tripling its training budget would increase this to 35%; this is still not reliable enough to win a coding contest! In contrast, engineering a system that samples from the model multiple times, tests each sample, etc. might increase performance to 80% with today’s models, as shown in work like AlphaCode. Even more importantly, iterating on a system design is often much faster than waiting for training runs. We believe that in any high-value application, developers will want to use every tool available to maximize AI quality, so they will use system ideas in addition to scaling. We frequently see this with LLM users, where a good LLM creates a compelling but frustratingly unreliable first demo, and engineering teams then go on to systematically raise quality.
Systems can be dynamic. Machine learning models are inherently limited because they are trained on static datasets, so their “knowledge” is fixed. Therefore, developers need to combine models with other components, such as search and retrieval, to incorporate timely data. In addition, training lets a model “see” the whole training set, so more complex systems are needed to build AI applications with access controls (e.g., answer a user’s questions based only on files the user has access to).
Improving control and trust is easier with systems. Neural network models alone are hard to control: while training will influence them, it is nearly impossible to guarantee that a model will avoid certain behaviors. Using an AI system instead of a model can help developers control behavior more tightly, e.g., by filtering model outputs. Likewise, even the best LLMs still hallucinate, but a system combining, say, LLMs with retrieval can increase user trust by providing citations or automatically verifying facts.
Performance goals vary widely. Each AI model has a fixed quality level and cost, but applications often need to vary these parameters. In some applications, such as inline code suggestions, the best AI models are too expensive, so tools like Github Copilot use carefully tuned smaller models and various search heuristics to provide results. In other applications, even the largest models, like GPT-4, are too cheap! Many users would be willing to pay a few dollars for a correct legal opinion, instead of the few cents it takes to ask GPT-4, but a developer would need to design an AI system to utilize this larger budget.
The shift to compound systems in Generative AI also matches the industry trends in other AI fields, such as self-driving cars: most of the state-of-the-art implementations are systems with multiple specialized components (more discussion here). For these reasons, we believe compound AI systems will remain a leading paradigm even as models improve.
Developing Compound AI Systems
While compound AI systems can offer clear benefits, the art of designing, optimizing, and operating them is still emerging. On the surface, an AI system is a combination of traditional software and AI models, but there are many interesting design questions. For example, should the overall “control logic” be written in traditional code (e.g., Python code that calls an LLM), or should it be driven by an AI model (e.g. LLM agents that call external tools)? Likewise, in a compound system, where should a developer invest resources—for example, in a RAG pipeline, is it better to spend more FLOPS on the retriever or the LLM, or even to call an LLM multiple times? Finally, how can we optimize an AI system with discrete components end-to-end to maximize a metric, the same way we can train a neural network? In this section, we detail a few example AI systems, then discuss these challenges and recent research on them.
The AI System Design Space
Below are few recent compound AI systems to show the breadth of design choices:
#mytable {
font-size: 16px;
}
#mytable ul {
font-size: 16px;
text-align: left;
}
#mytable td {
vertical-align: top;
}
#mytable th {
font-weight: bold;
}
#mytable ul {
padding-left: 15px;
}
AI System
Components
Design
Results
AlphaCode 2
Fine-tuned LLMs for sampling and scoring programs
Code execution module
Clustering model
Generates up to 1 million solutions for a coding problem then filters and scores them
Matches 85th percentile of humans on coding contests
AlphaGeometry
Fine-tuned LLM
Symbolic math engine
Iteratively suggests constructions in a geometry problem via LLM and checks deduced facts produced by symbolic engine
Between silver and gold International Math Olympiad medalists on timed test
Medprompt
GPT-4 LLM
Nearest-neighbor search in database of correct examples
LLM-generated chain-of-thought examples
Multiple samples and ensembling
Answers medical questions by searching for similar examples to construct a few-shot prompt, adding model-generated chain-of-thought for each example, and generating and judging up to 11 solutions
Outperforms specialized medical models like Med-PaLM used with simpler prompting strategies
Gemini on MMLU
Gemini LLM
Custom inference logic
Gemini's CoT@32 inference strategy for the MMLU benchmark samples 32 chain-of-thought answers from the model, and returns the top choice if enough of them agree, or uses generation without chain-of-thought if not
90.04% on MMLU, compared to 86.4% for GPT-4 with 5-shot prompting or 83.7% for Gemini with 5-shot prompting
ChatGPT Plus
LLM
Web Browser plugin for retrieving timely content
Code Interpreter plugin for executing Python
DALL-E image generator
The ChatGPT Plus offering can call tools such as web browsing to answer questions; the LLM determines when and how to call each tool as it responds
Popular consumer AI product with millions of paid subscribers
RAG,
ORQA,
Bing,
Baleen, etc
LLM (sometimes called multiple times)
Retrieval system
Combine LLMs with retrieval systems in various ways, e.g., asking an LLM to generate a search query, or directly searching for the current context
Widely used technique in search engines and enterprise apps
Key Challenges in Compound AI Systems
Compound AI systems pose new challenges in design, optimization and operation compared to AI models.
Design Space
The range of possible system designs for a given task is vast. For example, even in the simple case of retrieval-augmented generation (RAG) with a retriever and language model, there are: (i) many retrieval and language models to choose from, (ii) other techniques to improve retrieval quality, such as query expansion or reranking models, and (iii) techniques to improve the LLM’s generated output (e.g., running another LLM to check that the output relates to the retrieved passages). Developers have to explore this vast space to find a good design.
In addition, developers need to allocate limited resources, like latency and cost budgets, among the system components. For example, if you want to answer RAG questions in 100 milliseconds, should you budget to spend 20 ms on the retriever and 80 on the LLM, or the other way around?
Optimization
Often in ML, maximizing the quality of a compound system requires co-optimizing the components to work well together. For example, consider a simple RAG application where an LLM sees a user question, generates a search query to send to a retriever, and then generates an answer. Ideally, the LLM would be tuned to generate queries that work well for that particular retriever, and the retriever would be tuned to prefer answers that work well for that LLM.
In single model development a la PyTorch, users can easily optimize a model end-to-end because the whole model is differentiable. However, compound AI systems contain non-differentiable components like search engines or code interpreters, and thus require new methods of optimization. Optimizing these compound AI systems is still a new research area; for example, DSPy offers a general optimizer for pipelines of pretrained LLMs and other components, while others systems, like LaMDA, Toolformer and AlphaGeometry, use tool calls during model training to optimize models for those tools.
Operation
Machine learning operations (MLOps) become more challenging for compound AI systems. For example, while it is easy to track success rates for a traditional ML model like a spam classifier, how should developers track and debug the performance of an LLM agent for the same task, which might use a variable number of “reflection” steps or external API calls to classify a message? We believe that a new generation of MLOps tools will be developed to tackle these problems. Interesting problems include:
Monitoring: How can developers most efficiently log, analyze, and debug traces from complex AI systems?
DataOps: Because many AI systems involve data serving components like vector DBs, and their behavior depends on the quality of data served, any focus on operations for these systems should additionally span data pipelines.
Security: Research has shown that compound AI systems, such as an LLM chatbot with a content filter, can create unforeseen security risks compared to individual models. New tools will be required to secure these systems.
Emerging Paradigms
To tackle the challenges of building compound AI systems, multiple new approaches are arising in the industry and in research. We highlight a few of the most widely used ones and examples from our research on tackling these challenges.
Designing AI Systems: Composition Frameworks and Strategies. Many developers are now using “language model programming” frameworks that let them build applications out of multiple calls to AI models and other components. These include component libraries like LangChain and LlamaIndex that developers call from traditional programs, agent frameworks like AutoGPT and BabyAGI that let an LLM drive the application, and tools for controlling LM outputs, like Guardrails, Outlines, LMQL and SGLang. In parallel, researchers are developing numerous new inference strategies to generate better outputs using calls to models and tools, such as chain-of-thought, self-consistency, WikiChat, RAG and others.
Automatically Optimizing Quality: DSPy. Coming from academia, DSPy is the first framework that aims to optimize a system composed of LLM calls and other tools to maximize a target metric. Users write an application out of calls to LLMs and other tools, and provide a target metric such as accuracy on a validation set, and then DSPy automatically tunes the pipeline by creating prompt instructions, few-shot examples, and other parameter choices for each module to maximize end-to-end performance. The effect is similar to end-to-end optimization of a multi-layer neural network in PyTorch, except that the modules in DSPy are not always differentiable layers. To do that, DSPy leverages the linguistic abilities of LLMs in a clean way: to specify each module, users write a natural language signature, such as user_question -> search_query, where the names of the input and output fields are meaningful, and DSPy automatically turns this into suitable prompts with instructions, few-shot examples, or even weight updates to the underlying language models.
Optimizing Cost: FrugalGPT and AI Gateways. The wide range of AI models and services available makes it challenging to pick the right one for an application. Moreover, different models may perform better on different inputs. FrugalGPT is a framework to automatically route inputs to different AI model cascades to maximize quality subject to a target budget. Based on a small set of examples, it learns a routing strategy that can outperform the best LLM services by up to 4% at the same cost, or reduce cost by up to 90% while matching their quality. FrugalGPT is an example of a broader emerging concept of AI gateways or routers, implemented in software like Databricks AI Gateway, OpenRouter, and Martian, to optimize the performance of each component of an AI application. These systems work even better when an AI task is broken into smaller modular steps in a compound system, and the gateway can optimize routing separately for each step.
Operation: LLMOps and DataOps. AI applications have always required careful monitoring of both model outputs and data pipelines to run reliably. With compound AI systems, however, the behavior of the system on each input can be considerably more complex, so it is important to track all the steps taken by the application and intermediate outputs. Software like LangSmith, Phoenix Traces, and Databricks Inference Tables can track, visualize and evaluate these outputs at a fine granularity, in some cases also correlating them with data pipeline quality and downstream metrics. In the research world, DSPy Assertions seeks to leverage feedback from monitoring checks directly in AI systems to improve outputs, and AI-based quality evaluation methods like MT-Bench, FAVA and ARES aim to automate quality monitoring.
Conclusion
Generative AI has excited every developer by unlocking a wide range of capabilities through natural language prompting. As developers aim to move beyond demos and maximize the quality of their AI applications, however, they are increasingly turning to compound AI systems as a natural way to control and enhance the capabilities of LLMs. Figuring out the best practices for developing compound AI systems is still an open question, but there are already exciting approaches to aid with design, end-to-end optimization, and operation. We believe that compound AI systems will remain the best way to maximize the quality and reliability of AI applications going forward, and may be one of the most important trends in AI in 2024.
BibTex for this post:
@misc{compound-ai-blog,
title={The Shift from Models to Compound AI Systems},
author={Matei Zaharia and Omar Khattab and Lingjiao Chen and Jared Quincy Davis
and Heather Miller and Chris Potts and James Zou and Michael Carbin
and Jonathan Frankle and Naveen Rao and Ali Ghodsi},
howpublished={\url{https://bair.berkeley.edu/blog/2024/02/18/compound-ai-systems/}},
year={2024}
}
The model delivers dramatically enhanced performance, with a breakthrough in long-context understanding across modalities.
The model delivers dramatically enhanced performance, with a breakthrough in long-context understanding across modalities.
The model delivers dramatically enhanced performance, with a breakthrough in long-context understanding across modalities.
The model delivers dramatically enhanced performance, with a breakthrough in long-context understanding across modalities.
The model delivers dramatically enhanced performance, with a breakthrough in long-context understanding across modalities.
The model delivers dramatically enhanced performance, with a breakthrough in long-context understanding across modalities.
We're bringing Gemini to more Google products
We're bringing Gemini to more Google products
We're bringing Gemini to more Google products
We're bringing Gemini to more Google products
We're bringing Gemini to more Google products
We're bringing Gemini to more Google products
Advancing AI reasoning in mathematics