Gemini Diffusion
Our state-of-the-art, experimental text diffusion model
Large-language models are the foundation of generative AI today. We’re using a technique called diffusion to explore a new kind of language model that gives users greater control, creativity, and speed in text generation.
What is a diffusion model?
Traditional autoregressive language models generate text one word – or token – at a time. This sequential process can be slow, and limit the quality and coherence of the output.
Diffusion models work differently. Instead of predicting text directly, they learn to generate outputs by refining noise, step-by-step. This means they can iterate on a solution very quickly and error correct during the generation process. This helps them excel at tasks like editing, including in the context of math and code.
Capabilities
-
Rapid response
Generates content significantly faster than even our fastest model so far.
-
More coherent text
Generates entire blocks of tokens at once, meaning it responds more coherently to a user’s prompt than autoregressive models.
-
Iterative refinement
Corrects errors during generation for more consistent outputs.
Benchmarks
Gemini Diffusion’s external benchmark performance is comparable to much larger models, whilst also being faster.
Benchmark |
Gemini Diffusion
|
Gemini 2.0 Flash-Lite
|
---|---|---|
Code
LiveCodeBench (v6)
|
30.9% | 28.5% |
Code
BigCodeBench
|
45.4% | 45.8% |
Code
LBPP (v2)
|
56.8% | 56.0% |
Code
SWE-Bench Verified*
|
22.9% | 28.5% |
Code
HumanEval
|
89.6% | 90.2% |
Code
MBPP
|
76.0% | 75.8% |
Science
GPQA Diamond
|
40.4% | 56.5% |
Mathematics
AIME 2025
|
23.3% | 20.0% |
Reasoning
BIG-Bench Extra Hard
|
15.0% | 21.0% |
Multilingual
Global MMLU (Lite)
|
69.1% | 79.0% |
Gemini Diffusion speed
Sampling speed excluding overhead
|
1479 tokens / sec |
Overhead
|
0.84 sec |
Try Gemini Diffusion
Gemini Diffusion is currently available as an experimental demo to help develop and refine future models. If you're interested in getting access to the demo, please sign up to the waitlist.