Mercury 2
Mercury 2 is the world's fastest reasoning language model, delivering over 1,000 tokens per second with diffusion-based parallel generation for instant, production-grade AI. Achieve superior intelligence at a fraction of the cost and latency of traditional models, perfect for agentic workflows, real-time voice, and scalable inference.
Available for Chat, Vision, and File Uploads.
Performance Benchmarks
How do you want to interact?
Start a Conversation
Ask anything.
Have a natural conversation, brainstorm ideas, draft emails, or ask for advice.
Use a Persona
Specialized Experts.
Instruct the AI to act as a Coding Tutor, Marketing Expert, or Travel Guide.
Why use Mercury 2?
Ultra-Fast Reasoning
Generates text at 1,000+ tokens per second using diffusion-based parallel refinement, 5x faster than leading models
Tunable Reasoning
Supports adjustable reasoning levels with step-by-step thinking for agentic tasks and complex problem-solving
Native Tool Use
Includes 128K context window, tool integration, and schema-aligned JSON output for production workflows
Capability Examples
Mercury 2 Speed Demo
Mercury 2 Reasoning Demo
How to use
Go to Chat
Navigate to the "AI Chat" page.
Select Model
Ensure Mercury 2 is selected.
Type Prompt
Ask a question or paste code.
Interact
Refine the answer by replying to the AI.
Compare LLMs Side-by-Side
Is Mercury 2 better than Claude 3.5 or Gemini? Test same prompts simultaneously in the Chat Playground.
Open Chat PlaygroundMade with ❤ by AI4Chat