Models
Inception’s diffusion LLMs (dLLMs) deliver frontier LLM quality at 5x greater speed.
Overview
Seamless integration
Our models are OpenAI compatible and a drop-in replacement for traditional LLMs.
Discover our models
Mercury 2
Pricing
Input
$0.25 / 1M Tokens
Cached Input
$0.025 / 1M Tokens
Output
$0.75 / 1M Tokens
Features
128K context window
Reasoning
Tool use
Structured Output
Use cases
Rapid Coding Iteration
Workflow Subagents
Customer Support
Realtime Voice
Enterprise Search
Mercury Edit
Pricing
Input
$0.25 / 1M Tokens
Cached Input
$0.025 / 1M Tokens
Output
$0.75 / 1M Tokens
Features
128K context window
Tool use
Structured Output
Use cases
Autocomplete
Next Edit
* Mercury 1 remains supported for existing customers. For access or migration guidance, contact your Inception representative.
Get started with Mercury today
1
Create your account
Create an Inception Platform account or sign in directly if you already have one.
2
Create your API Key
Go to API Keys and create a new API key. New API keys comes with 10 million free tokens
3
Make your first request
We are OpenAI API compatible and are supported through libraries including AISuite, LiteLLM, and LangChain.
Pricing
Choose the access plan that works best for your needs
Free
For people that want to try out our playground
Access all models
10 million free tokens
Developer
For people that want to try out our playground
Usage-based pricing
Generous rate limits
Priority support
Enterprise
For people that want to try out our playground
Custom rate limits
SLA guarantees
Security and privacy
Volume-based pricing

