Mercury 2

Inception mercury-2
Model Information
Slug mercury-2
LLM.txt View
Release Date February 24, 2026
AIME 2025 0.911
GPQA 0.74
Organization
Model Description
Mercury 2 is an extremely fast reasoning LLM, and the first reasoning diffusion LLM (dLLM).
Instead of generating tokens sequentially, Mercury 2 produces and refines multiple tokens in parallel, achieving >1,000 tokens/sec on standard GPUs. Mercury 2 is 5x+ faster than leading speed-optimized LLMs like Claude 4.5 Haiku and GPT 5 Mini, at a fraction of the cost.
Mercury 2 supports tunable reasoning levels, 128K context, native tool use, and schema-aligned JSON output. Built for coding workflows where latency compounds, real-time voice/search, and agent loops. OpenAI API compatible. Read more in the [blog post](https://www.inceptionlabs.ai/blog/introducing-mercury-2).
Available at 12 Providers
Provider Type Model Name Original Model Input ($/1M) Output ($/1M) Free Actions
Routeway
Routeway
Inception Labs: Mercury 2
mercury-2 $0.23 $0.69
Kilo Code
Kilo Code
Code
Inception: Mercury 2
inception/mercury-2 $0.25 $0.75
OpenRouter
OpenRouter
Chat Code
Mercury 2
inception/mercury-2 $0.25 $0.75
Inception
Inception
Mercury 2
mercury-2 $0.25 $0.75
Venice
Venice
Mercury 2
mercury-2 $0.31 $0.94
Nano-GPT
Nano-GPT
Mercury 2
mercury-2 - -
Yupp
Yupp
Chat
Mercury 2 (OpenRouter)
inception/mercury-2 - -
Writingmate
Writingmate
Chat Code
Inception: Mercury 2
inception/mercury-2 - -
Arena AI
Arena AI
Chat
mercury-2 - -
ValorGPT
ValorGPT
IN
inception-mercury-2 - -
LLM Stats
LLM Stats
Chat
Mercury 2
mercury-2 - -
LangDB
LangDB
mercury-2
mercury-2 - -