Inception: Mercury Coder
INCEPTION Developer Architecture Profile
Intelligence (ELO)1120Chatbot Arena Verified
Max Context128,000Tokens
API Cost / 1M$1.00Blended Prompt + Completion
Model Capabilities
- Coding & Logic
Mercury Coder is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like Claude 3.5 Haiku and GPT-4o Mini while matching their performance. Mercury Coder's speed means that developers can stay in the flow while coding, enjoying rapid chat-based iteration and responsive code completion suggestions. On Copilot Arena, Mercury Coder ranks 1st in speed and ties for 2nd in quality. Read more in the [blog post here](https://www.inceptionlabs.ai/blog/introducing-mercury).
Granular Pricing Matrix
Input Tokens (Prompt)$0.25 / 1M
Output Tokens (Completion)$0.75 / 1M
Pricing data via OpenRouter. Sync: 3/16/2026
Evaluate Competitors
VS Engine MatchupInception: Mercury Coder vs Z.ai: GLM 5 TurboVS Engine MatchupInception: Mercury Coder vs Inception: Mercury 2VS Engine MatchupInception: Mercury Coder vs Qwen: Qwen3.5-27BVS Engine MatchupInception: Mercury Coder vs Qwen: Qwen3.5-122B-A10BVS Engine MatchupInception: Mercury Coder vs AionLabs: Aion-2.0VS Engine MatchupInception: Mercury Coder vs Qwen: Qwen3.5 Plus 2026-02-15