Directory
Model directory
270 modelsDiscover frontier APIs, open-weight models, and enterprise-ready systems using filters built for analysts and buyers. Full-profile entries appear alongside verified listings so breadth does not come at the cost of freshness, provenance, or confidence signaling.
Last verified dates and source confidence stay attached to every model card and detail page. Models with incomplete public data remain listed for market visibility, but rankings and compare views prioritize full-profile entries over verified listings.
Entries eligible for rankings and the strongest comparison claims.
Source-backed records that preserve access details and primary links even when benchmark coverage is sparse.
Search spans provider, family, modalities, summary text, access modes, and freshness-aware catalog data.
OpenAI
GPT-5.4
OpenAI
OpenAI's GPT-5.4, the most capable and efficient frontier model for professional work. First general-purpose model with native computer-use capabilities. Combines industry-leading coding from GPT-5.3-Codex with improved agentic workflows.
- Context
- 1,000,000
- Input
- $0.005/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Anthropic
Claude Sonnet 4.6
Claude 4.6
Anthropic's current Sonnet tier for fast frontier reasoning, coding, and long-context agent work.
- Context
- 1,000,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Anthropic
Claude Opus 4.6
Claude 1M
Anthropic's most intelligent Claude model for complex agents, coding, and deep reasoning, with 1M token context and 128K output.
- Context
- 1,000,000
- Input
- $0.005/1K tok
- Output
- $0.03/1K tok
- Coverage
- Full profile
Anthropic
Claude Sonnet 4.5
Claude 1M
Anthropic's Sonnet 4.5 with 1M token context for fast frontier reasoning, coding, and long-context agent work.
- Context
- 1,000,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 3.1 Pro
Gemini 3.1
Google's Gemini 3.1 Pro, designed for complex tasks where simple answers aren't enough. Released Feb 2026 with enhanced reasoning and multimodal capabilities.
- Context
- 1,048,576
- Input
- $0.0013/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
OpenAI
GPT-4o
GPT-4
A broadly capable multimodal model optimized for production chat, agentic workflows, and voice experiences.
- Context
- 128,000
- Input
- $0.005/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
OpenAI
GPT-5.2
OpenAI
OpenAI's latest frontier model with 1M token context for advanced reasoning and tool use.
- Context
- 1,000,000
- Input
- $0.005/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
OpenAI
GPT-5.4 Pro
OpenAI
OpenAI's GPT-5.4 Pro delivering maximum performance for complex tasks. Available for ChatGPT Pro and Enterprise plans.
- Context
- 1,000,000
- Input
- $0.01/1K tok
- Output
- $0.03/1K tok
- Coverage
- Full profile
Anthropic
Claude 3.7 Sonnet
Claude 3.7
A top-tier reasoning model with strong software engineering assistance and enterprise controls.
- Context
- 200,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 2.5 Pro
Gemini 2.5
An expansive-context model built for complex research, multimodal understanding, and large document tasks, with 1M token native context.
- Context
- 1,048,576
- Input
- $0.0013/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
OpenAI
GPT-5.2 Pro
OpenAI
OpenAI's premium GPT-5.2 Pro with maximum reasoning capability.
- Context
- 1,000,000
- Input
- $0.01/1K tok
- Output
- $0.03/1K tok
- Coverage
- Full profile
Anthropic
Claude Opus 4.1
Claude 1M
Anthropic's Claude Opus 4.1 for maximum intelligence on complex reasoning, coding, and agentic tasks.
- Context
- 200,000
- Input
- $0.02/1K tok
- Output
- $0.08/1K tok
- Coverage
- Full profile
Anthropic
Claude Sonnet 4
Claude 1M
Anthropic's Sonnet 4 with 1M token context for balanced reasoning and coding performance.
- Context
- 1,000,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 3.0 Pro
Gemini 3.0
Google's Gemini 3.0 Pro, quietly released in late 2025. Strong multimodal reasoning with 1M context. Foundation for the Gemini 3.1 series.
- Context
- 1,048,576
- Input
- $0.0013/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 3.1 Flash
Gemini 3.1
Google's Gemini 3.1 Flash for fast, cost-efficient multimodal inference with strong quality.
- Context
- 1,048,576
- Input
- $0.0002/1K tok
- Output
- $0.0006/1K tok
- Coverage
- Full profile
OpenAI
GPT-5
OpenAI
OpenAI's GPT-5 with 1M context for balanced reasoning and coding.
- Context
- 1,000,000
- Input
- $0.005/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 2.5 Pro TTS
Gemini
Google's Gemini 2.5 Pro with 1M context for long-context multimodal analysis and text-to-speech.
- Context
- 1,048,576
- Input
- $0.0013/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
Anthropic
Claude Opus 4
Claude 200K
Anthropic's Opus 4 with 200K context for complex reasoning and agentic tasks.
- Context
- 200,000
- Input
- $0.02/1K tok
- Output
- $0.08/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 2.5 Flash
Gemini
Google's Gemini 2.5 Flash with 1M context for fast, cost-efficient multimodal inference.
- Context
- 1,048,576
- Input
- $0.0002/1K tok
- Output
- $0.0006/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 3.0 Flash
Gemini 3.0
Google's Gemini 3.0 Flash for fast multimodal inference. Released alongside Gemini 3.0 Pro.
- Context
- 1,048,576
- Input
- $0.0002/1K tok
- Output
- $0.0006/1K tok
- Coverage
- Full profile
OpenAI
GPT-5.3-Codex
OpenAI
OpenAI's GPT-5.3-Codex for expanding Codex across the full spectrum of professional work on a computer. Powers the Codex app on macOS and Windows.
- Context
- 1,000,000
- Input
- $0.005/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
OpenAI
GPT-5 mini
OpenAI
OpenAI's GPT-5 mini for cost-efficient applications.
- Context
- 1,000,000
- Input
- $0.0015/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Mistral AI
Mistral Large 25
Mistral Large
An enterprise-ready all-rounder with especially strong European language performance.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 2.0 Flash
Gemini
Google's Gemini 2.0 Flash with 1M context for fast multimodal inference.
- Context
- 1,048,576
- Input
- $0.0001/1K tok
- Output
- $0.0004/1K tok
- Coverage
- Full profile
OpenAI
GPT-4.1
OpenAI
OpenAI's GPT-4.1 with 1M token context, improved coding and instruction following.
- Context
- 1,048,576
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
OpenAI
GPT-5.3 Instant
OpenAI
OpenAI's GPT-5.3 Instant optimized for low-latency everyday usability. Available to all ChatGPT users.
- Context
- 1,000,000
- Input
- $0.0005/1K tok
- Output
- $0.002/1K tok
- Coverage
- Full profile
Anthropic
Claude Haiku 4.5
Claude 200K
Anthropic's Haiku 4.5 with 200K context, the fastest Claude model with near-frontier intelligence at low cost.
- Context
- 200,000
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
Meta
Llama 4 Maverick
Llama 4
Meta's 17Bx128E MoE open-weight model with 1M token context, pretrained on ~22T tokens. Strong multimodal and multilingual capabilities for teams that need control, private deployment, and customization.
- Context
- 1,048,576
- Input
- $0.0008/1K tok
- Output
- $0.002/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 2.5 Flash Live
Gemini
Google's lightweight Gemini 2.5 Flash variants for cost-efficient multimodal and real-time audio.
- Context
- 1,048,576
- Input
- $0.0001/1K tok
- Output
- $0.0003/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 2.5 Flash Native Audio Preview
Gemini
Google's lightweight Gemini 2.5 Flash variants for cost-efficient multimodal and real-time audio.
- Context
- 1,048,576
- Input
- $0.0001/1K tok
- Output
- $0.0003/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 2.5 Flash-Lite
Gemini
Google's lightweight Gemini 2.5 Flash variants for cost-efficient multimodal and real-time audio.
- Context
- 1,048,576
- Input
- $0.0001/1K tok
- Output
- $0.0003/1K tok
- Coverage
- Full profile
Cohere
Command R+ 2026
Command R+
An enterprise retrieval specialist built for document-heavy support and knowledge applications.
- Context
- 128,000
- Input
- $0.0015/1K tok
- Output
- $0.004/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 1.5 Pro
Gemini
Google's Gemini 1.5 Pro with 2M context for long document and media analysis.
- Context
- 2,097,152
- Input
- $0.0013/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 3.1 Flash-Lite
Gemini 3.1
Google's Gemini 3.1 Flash-Lite, their most cost-effective AI model. Released Mar 2026 for high-volume workloads with best-in-class intelligence at scale.
- Context
- 1,048,576
- Input
- $0.0001/1K tok
- Output
- $0.0003/1K tok
- Coverage
- Full profile
OpenAI
o3
OpenAI
OpenAI's o3 reasoning model for complex multi-step problem solving with extended thinking.
- Context
- 200,000
- Input
- $0.01/1K tok
- Output
- $0.04/1K tok
- Coverage
- Full profile
OpenAI
o4-mini
OpenAI
OpenAI's o4-mini for cost-efficient reasoning with strong coding and math capabilities.
- Context
- 200,000
- Input
- $0.0011/1K tok
- Output
- $0.0044/1K tok
- Coverage
- Full profile
OpenAI
GPT-4.1 mini
OpenAI
OpenAI's GPT-4.1 mini for fast, cost-efficient inference with 1M context.
- Context
- 1,048,576
- Input
- $0.0004/1K tok
- Output
- $0.0016/1K tok
- Coverage
- Full profile
OpenAI
o3-deep-research
OpenAI
OpenAI's deep research model for extended multi-step research tasks with tool use.
- Context
- 200,000
- Input
- $0.01/1K tok
- Output
- $0.04/1K tok
- Coverage
- Full profile
OpenAI
o4-mini-deep-research
OpenAI
OpenAI's cost-efficient deep research model for extended tasks.
- Context
- 200,000
- Input
- $0.003/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
Anthropic
Claude Haiku 3.5
Claude 200K
Anthropic's Haiku 3.5 with 200K context for fast, cost-efficient inference.
- Context
- 200,000
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
Anthropic
Claude Opus 3
Claude 200K
Anthropic's legacy Opus 3 model with 200K context.
- Context
- 200,000
- Input
- $0.02/1K tok
- Output
- $0.08/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 2.0 Flash-Lite
Gemini
Google's Gemini 2.0 Flash-Lite for ultra-fast, cost-efficient inference.
- Context
- 1,048,576
- Input
- $0.0001/1K tok
- Output
- $0.0003/1K tok
- Coverage
- Full profile
OpenAI
GPT-5 nano
OpenAI
OpenAI's smallest GPT-5 variant for ultra-fast, cost-sensitive applications.
- Context
- 1,000,000
- Input
- $0.0005/1K tok
- Output
- $0.002/1K tok
- Coverage
- Full profile
Amazon Web Services
Nova Pro
Nova
Amazon's Nova Pro model on Bedrock with 300K context for enterprise text and vision tasks.
- Context
- 300,000
- Input
- $0.0008/1K tok
- Output
- $0.0032/1K tok
- Coverage
- Full profile
Anthropic
Claude Sonnet 3
Claude 200K
Anthropic's legacy Sonnet 3 model with 200K context.
- Context
- 200,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 1.5 Flash
Gemini
Google's Gemini 1.5 Flash with 1M context for fast multimodal inference.
- Context
- 1,048,576
- Input
- $0.0001/1K tok
- Output
- $0.0003/1K tok
- Coverage
- Full profile
OpenAI
GPT-4o-mini
OpenAI
OpenAI's GPT-4o-mini for fast, cost-efficient inference with strong general capability.
- Context
- 128,000
- Input
- $0.0002/1K tok
- Output
- $0.0006/1K tok
- Coverage
- Full profile
OpenAI
o1
OpenAI
OpenAI's o1 reasoning model with advanced chain-of-thought for complex problem solving.
- Context
- 200,000
- Input
- $0.02/1K tok
- Output
- $0.06/1K tok
- Coverage
- Full profile
Perplexity
Sonar Reasoning Pro
Sonar
Perplexity's Sonar Reasoning Pro for complex reasoning with real-time web search. $2/$8 per MTok plus request fee.
- Context
- 200,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Meta
Llama 3.3 70B Instruct
Llama
Meta's latest Llama 3.x dense model, 70B parameters with 128K context.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Amazon Web Services
Nova Lite
Nova
Amazon's Nova Lite model on Bedrock for fast, cost-efficient multimodal inference.
- Context
- 300,000
- Input
- $0.0001/1K tok
- Output
- $0.0002/1K tok
- Coverage
- Full profile
OpenAI
o1-mini
OpenAI
OpenAI's o1-mini for fast, cost-efficient reasoning on coding and STEM tasks.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
Perplexity
Sonar Pro
Sonar
Perplexity's Sonar Pro for enhanced web search with multi-step reasoning, Pro Search, and citation support. $3/$15 per MTok plus request fee.
- Context
- 200,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Google DeepMind
Gemini 1.5 Flash-8B
Gemini
Google's lightweight Gemini 1.5 Flash-8B for ultra-fast, cost-efficient inference.
- Context
- 1,048,576
- Input
- $0.00/1K tok
- Output
- $0.0002/1K tok
- Coverage
- Full profile
Meta
Llama 4 Scout
Llama
Meta's Llama 4 Scout (17Bx16E MoE, 109B total params) with an extraordinary 10M token context window.
- Context
- 10,485,760
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Anthropic
Claude Haiku 3
Claude 200K
Anthropic's deprecated Haiku 3 model, retiring April 2026.
- Context
- 200,000
- Input
- $0.0003/1K tok
- Output
- $0.0013/1K tok
- Coverage
- Full profile
Anthropic
Claude Haiku 3
Claude 200K
Anthropic's deprecated Haiku 3 model, retiring April 2026.
- Context
- 200,000
- Input
- $0.0003/1K tok
- Output
- $0.0013/1K tok
- Coverage
- Full profile
Anthropic
Claude Haiku 3.5
Claude 200K
Anthropic's Haiku 3.5 with 200K context for fast, cost-efficient inference.
- Context
- 200,000
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
Anthropic
Claude Haiku 4.5
Claude 200K
Anthropic's Haiku 4.5 with 200K context, the fastest Claude model with near-frontier intelligence at low cost.
- Context
- 200,000
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
Mistral AI
Codestral
Mistral audio / tooling
Mistral's cutting-edge code completion model with 256K context fill-in-the-middle support.
- Context
- 256,000
- Input
- $0.001/1K tok
- Output
- $0.003/1K tok
- Coverage
- Full profile
Mistral AI
Codestral Embed
Mistral audio / tooling
Mistral's embedding models for semantic search and code representation extraction.
- Context
- 32,768
- Input
- $0.001/1K tok
- Output
- $0.003/1K tok
- Coverage
- Full profile
Z.AI
CogVideoX
GLM multimodal
Z.AI's multimodal model line for document OCR, image generation, video generation, and visual reasoning tasks.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Z.AI
CogView 4
GLM multimodal
Z.AI's multimodal model line for document OCR, image generation, video generation, and visual reasoning tasks.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Mistral AI
Devstral Medium 1.0
Mistral
Mistral's enterprise-grade model excelling at software engineering tasks.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
ByteDance / Doubao
Doubao-Seed-1.6
Doubao
Bytedance's Doubao coding and reasoning line for domestic developer workflows, visual coding, and agentic assistance.
- Context
- 128,000
- Input
- $0.001/1K tok
- Output
- $0.004/1K tok
- Coverage
- Full profile
ByteDance / Doubao
Doubao-Seed-1.6-Flash
Doubao
Bytedance's Doubao coding and reasoning line for domestic developer workflows, visual coding, and agentic assistance.
- Context
- 128,000
- Input
- $0.001/1K tok
- Output
- $0.004/1K tok
- Coverage
- Full profile
ByteDance / Doubao
Doubao-Seed-2.0-Code
Doubao
Bytedance's Doubao coding and reasoning line for domestic developer workflows, visual coding, and agentic assistance.
- Context
- 128,000
- Input
- $0.001/1K tok
- Output
- $0.004/1K tok
- Coverage
- Full profile
ByteDance / Doubao
Doubao-Seed-Code
Doubao
Bytedance's Doubao coding and reasoning line for domestic developer workflows, visual coding, and agentic assistance.
- Context
- 128,000
- Input
- $0.001/1K tok
- Output
- $0.004/1K tok
- Coverage
- Full profile
Baidu / ERNIE
ERNIE 3.5 128K
ERNIE
Baidu's ERNIE family covers enterprise chat, long-context, speed-optimized, and function-calling workloads.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Baidu / ERNIE
ERNIE 4.0 Turbo 8K
ERNIE
Baidu's ERNIE family covers enterprise chat, long-context, speed-optimized, and function-calling workloads.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Baidu / ERNIE
ERNIE Functions 8K
ERNIE
Baidu's ERNIE family covers enterprise chat, long-context, speed-optimized, and function-calling workloads.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Baidu / ERNIE
ERNIE Speed 128K
ERNIE
Baidu's ERNIE family covers enterprise chat, long-context, speed-optimized, and function-calling workloads.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Z.AI
GLM-4.5
GLM
Z.AI's GLM reasoning family for general chat, coding, and tool-enabled Chinese-first production workloads.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Z.AI
GLM-4.5V
GLM multimodal
Z.AI's multimodal model line for document OCR, image generation, video generation, and visual reasoning tasks.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Z.AI
GLM-4.6
GLM
Z.AI's GLM reasoning family for general chat, coding, and tool-enabled Chinese-first production workloads.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Z.AI
GLM-4.6V
GLM multimodal
Z.AI's multimodal model line for document OCR, image generation, video generation, and visual reasoning tasks.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Z.AI
GLM-4.7
GLM
Z.AI's GLM reasoning family for general chat, coding, and tool-enabled Chinese-first production workloads.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Z.AI
GLM-5
GLM
Z.AI's GLM reasoning family for general chat, coding, and tool-enabled Chinese-first production workloads.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Z.AI
GLM-Image
GLM multimodal
Z.AI's multimodal model line for document OCR, image generation, video generation, and visual reasoning tasks.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Z.AI
GLM-OCR
GLM multimodal
Z.AI's multimodal model line for document OCR, image generation, video generation, and visual reasoning tasks.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
OpenAI
gpt-audio
OpenAI audio
OpenAI's realtime and audio models for low-latency voice interfaces.
- Context
- 128,000
- Input
- $0.006/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
OpenAI
gpt-oss-120b
gpt-oss
OpenAI's 120B open-weight model for frontier-style reasoning with self-hosted deployment.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
OpenAI
gpt-realtime
OpenAI audio
OpenAI's realtime and audio models for low-latency voice interfaces.
- Context
- 128,000
- Input
- $0.006/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Tencent / Hunyuan
Hunyuan Code
Hunyuan
Tencent's Hunyuan family spans fast-reasoning, long-context, coding, and vision-capable API deployments.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Tencent / Hunyuan
Hunyuan Lite
Hunyuan
Tencent's Hunyuan family spans fast-reasoning, long-context, coding, and vision-capable API deployments.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Tencent / Hunyuan
Hunyuan Standard
Hunyuan
Tencent's Hunyuan family spans fast-reasoning, long-context, coding, and vision-capable API deployments.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Tencent / Hunyuan
Hunyuan T1
Hunyuan
Tencent's Hunyuan T1 flagship with 256K context for fast reasoning and long-document understanding.
- Context
- 256,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Tencent / Hunyuan
Hunyuan T1 Vision
Hunyuan
Tencent's Hunyuan family spans fast-reasoning, long-context, coding, and vision-capable API deployments.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Tencent / Hunyuan
Hunyuan TurboS
Hunyuan
Tencent's Hunyuan family spans fast-reasoning, long-context, coding, and vision-capable API deployments.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Tencent / Hunyuan
Hunyuan TurboS LongText 128K
Hunyuan
Tencent's Hunyuan family spans fast-reasoning, long-context, coding, and vision-capable API deployments.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Moonshot AI / Kimi
Kimi K2
Kimi
Moonshot AI's Kimi K2 with 128K native context (extensible to 131K) for research, code generation, and reasoning-heavy chat.
- Context
- 131,072
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Moonshot AI / Kimi
Kimi K2 Thinking
Kimi
Moonshot AI's Kimi family variants with extended context and thinking modes for complex reasoning tasks.
- Context
- 256,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Moonshot AI / Kimi
Kimi K2 Turbo Preview
Kimi
Moonshot AI's Kimi family variants with extended context and thinking modes for complex reasoning tasks.
- Context
- 256,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Moonshot AI / Kimi
Kimi K2.5
Kimi
Moonshot AI's Kimi family variants with extended context and thinking modes for complex reasoning tasks.
- Context
- 256,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Mistral AI
Magistral Medium 1.2
Mistral
Mistral's frontier-class multimodal reasoning model with extended thinking.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
MiniMax
MiniMax-M1
MiniMax media
MiniMax's text generation models with 200K context for general-purpose language tasks.
- Context
- 204,800
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
MiniMax
MiniMax-M2
MiniMax
MiniMax's earlier M2.x models for general-purpose multimodal inference.
- Context
- 204,800
- Input
- $0.002/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
MiniMax
MiniMax-M2.1
MiniMax
MiniMax's earlier M2.x models for general-purpose multimodal inference.
- Context
- 204,800
- Input
- $0.002/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
MiniMax
MiniMax-M2.1-highspeed
MiniMax
MiniMax's earlier M2.x models for general-purpose multimodal inference.
- Context
- 204,800
- Input
- $0.002/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
MiniMax
MiniMax-M2.5
MiniMax
MiniMax's latest M2.5 text model for coding agents, multimodal assistants, and high-speed inference.
- Context
- 204,800
- Input
- $0.002/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
MiniMax
MiniMax-M2.5-highspeed
MiniMax
MiniMax's latest M2.5 text model for coding agents, multimodal assistants, and high-speed inference.
- Context
- 204,800
- Input
- $0.002/1K tok
- Output
- $0.01/1K tok
- Coverage
- Full profile
MiniMax
MiniMax-Text-01
MiniMax media
MiniMax's text generation models with 200K context for general-purpose language tasks.
- Context
- 204,800
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
MiniMax
MiniMax-VL-01
MiniMax media
MiniMax's vision-language model with 200K context for multimodal understanding and image analysis.
- Context
- 204,800
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
Mistral AI
Mistral Embed
Mistral audio / tooling
Mistral's embedding models for semantic search and code representation extraction.
- Context
- 32,768
- Input
- $0.001/1K tok
- Output
- $0.003/1K tok
- Coverage
- Full profile
Mistral AI
Mistral Large 3
Mistral
Mistral's state-of-the-art open-weight multimodal model with 128K context for general-purpose deployment.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Mistral AI
Mistral Medium 3.1
Mistral
Mistral's frontier-class multimodal model released August 2025.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Mistral AI
Mistral Moderation
Mistral audio / tooling
Mistral's moderation model for detecting harmful text content and policy violations.
- Context
- 32,768
- Input
- $0.001/1K tok
- Output
- $0.003/1K tok
- Coverage
- Full profile
Mistral AI
Mistral Small 3.1
Mistral
Mistral's efficient small models with 128K context, unifying instruct, reasoning, and coding.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Mistral AI
Mistral Small 3.2 Open
Mistral
Mistral's efficient small models with 128K context, unifying instruct, reasoning, and coding.
- Context
- 128,000
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Mistral AI
Pixtral 12B
Mistral vision / audio
Mistral's 12B open-weight vision model for private document analysis and on-prem image understanding.
- Context
- 131,072
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Mistral AI
Pixtral Large
Mistral vision / audio
Mistral's first frontier-class multimodal model with 128K context for image understanding and document analysis.
- Context
- 131,072
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Perplexity
Sonar Deep Research
Sonar
Perplexity's Sonar Deep Research for automated multi-step research with citation generation. $2/$8 MTok + $2 citation + $3 reasoning + $5 per 1K queries.
- Context
- 200,000
- Input
- $0.002/1K tok
- Output
- $0.008/1K tok
- Coverage
- Full profile
Z.AI
Vidu Q1
GLM multimodal
Z.AI's multimodal model line for document OCR, image generation, video generation, and visual reasoning tasks.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Mistral AI
Voxtral Mini Open
Mistral vision / audio
Mistral's open-weight mini audio input model for on-prem transcription and audio understanding.
- Context
- 131,072
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Mistral AI
Voxtral Mini Transcribe
Mistral audio / tooling
Mistral's audio transcription model optimized for converting speech to text with high accuracy.
- Context
- 131,072
- Input
- $0.001/1K tok
- Output
- $0.003/1K tok
- Coverage
- Full profile
Mistral AI
Voxtral Small Open
Mistral vision / audio
Mistral's open-weight small audio input model for instruct use cases with audio understanding.
- Context
- 131,072
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Anthropic
Claude Sonnet 3
Claude 200K
Anthropic's legacy Sonnet 3 model with 200K context.
- Context
- 200,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Anthropic
Claude Sonnet 4
Claude 1M
Anthropic's Sonnet 4 with 1M token context for balanced reasoning and coding performance.
- Context
- 1,000,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Cohere
Command A
Command
Cohere's latest Command model with 256K context for enterprise chat, retrieval, and agent workflows.
- Context
- 256,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Cohere
Command A Reasoning
Command
Cohere's Command text family for enterprise chat, retrieval, translation, and agent workflows built around grounding.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Cohere
Command A Translate
Command
Cohere's Command text family for enterprise chat, retrieval, translation, and agent workflows built around grounding.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Cohere
Command A Vision
Command
Cohere's vision-capable Command and embedding models for multimodal enterprise assistants and retrieval systems.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Cohere
Command R+
Command
Cohere's Command text family for enterprise chat, retrieval, translation, and agent workflows built around grounding.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Cohere
Command R7B
Command
Cohere's Command text family for enterprise chat, retrieval, translation, and agent workflows built around grounding.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
DeepSeek
DeepSeek-Coder-V2
DeepSeek Coder/Math
DeepSeek's specialist models for code generation and mathematical reasoning.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-Math-V2
DeepSeek Coder/Math
DeepSeek's specialist models for code generation and mathematical reasoning.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-R1
DeepSeek R1
DeepSeek's reasoning model (671B total, 37B activated) trained with large-scale RL, achieving o1-level performance on math, code, and reasoning tasks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-R1-Distill-Llama-70B
DeepSeek distill
DeepSeek-R1 distilled into Llama 3.3 70B, achieving strong reasoning with dense model efficiency.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V2.5
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V3
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V3.1
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V3.1-Base
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V3.2
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V3.2-Exp
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Mistral AI
Devstral 2 Open
Mistral open
Mistral's open-weight reasoning and coding models for self-hosted deployment.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Mistral AI
Devstral Small 2
Mistral open
Mistral's open-weight reasoning and coding models for self-hosted deployment.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Cohere
Embed 4
Command
Cohere's vision-capable Command and embedding models for multimodal enterprise assistants and retrieval systems.
- Context
- 128,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Baidu / ERNIE
ERNIE 4.5 Turbo 32K
ERNIE
Baidu's ERNIE 4.5 Turbo with 32K context for mainstream enterprise chat and function-calling workloads.
- Context
- 32,768
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
xAI
Grok 3
Grok
xAI's Grok API family for fast-moving reasoning, conversational agents, and multimodal assistant workloads.
- Context
- 131,072
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
xAI
Grok 3 Mini
Grok
xAI's Grok API family for fast-moving reasoning, conversational agents, and multimodal assistant workloads.
- Context
- 131,072
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
xAI
Grok 4
Grok
xAI's latest Grok model with 256K context for advanced reasoning and multimodal tasks.
- Context
- 256,000
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
xAI
Grok 4 Fast Reasoning
Grok
xAI's Grok API family for fast-moving reasoning, conversational agents, and multimodal assistant workloads.
- Context
- 131,072
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
xAI
grok-image
Grok
xAI's Grok API family for fast-moving reasoning, conversational agents, and multimodal assistant workloads.
- Context
- 131,072
- Input
- $0.003/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
AI21 Labs
Jamba 3B
Jamba
AI21's earlier Jamba models for long-context enterprise assistants with open deployment options.
- Context
- 256,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
AI21 Labs
Jamba Large
Jamba
AI21's earlier Jamba models for long-context enterprise assistants with open deployment options.
- Context
- 256,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
AI21 Labs
Jamba Large 1.6
Jamba
AI21's flagship Jamba model with 256K context for long-document enterprise workflows.
- Context
- 256,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
AI21 Labs
Jamba Mini
Jamba
AI21's earlier Jamba models for long-context enterprise assistants with open deployment options.
- Context
- 256,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
AI21 Labs
Jamba Mini 1.6
Jamba
AI21's efficient Jamba Mini models with 256K context for cost-sensitive long-context deployments.
- Context
- 256,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
AI21 Labs
Jamba Mini 1.7
Jamba
AI21's efficient Jamba Mini models with 256K context for cost-sensitive long-context deployments.
- Context
- 256,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Mistral AI
Magistral Small 1.2 Open
Mistral open
Mistral's open-weight reasoning and coding models for self-hosted deployment.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Mistral AI
Ministral 3 14B Open
Mistral open
Mistral's efficient open-weight models with text and vision capabilities at 3B, 8B, and 14B sizes.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Mistral AI
Ministral 3 3B Open
Mistral open
Mistral's efficient open-weight models with text and vision capabilities at 3B, 8B, and 14B sizes.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Mistral AI
Ministral 3 8B Open
Mistral open
Mistral's efficient open-weight models with text and vision capabilities at 3B, 8B, and 14B sizes.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Mistral AI
Mistral Large 3 Open
Mistral open
Mistral's state-of-the-art open-weight model with 128K context for self-hosting and customization.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Mistral AI
Mistral Nemo 12B
Mistral open
Mistral's open-weight reasoning and coding models for self-hosted deployment.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Mistral AI
Mistral OCR 2505
Mistral vision / audio
Mistral's OCR model for document extraction, PDF parsing, and structured output from images.
- Context
- 32,768
- Input
- $0.002/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Morph
morph-v3-fast-apply
Morph Fast Apply
Morph's Fast Apply model merges AI-generated code edits at 10,500+ tokens/sec with 98% accuracy. 10x faster than alternatives like GPT-4.1 (55 tok/s) or Claude Sonnet (80 tok/s).
- Context
- 128,000
- Input
- $0.0008/1K tok
- Output
- $0.002/1K tok
- Coverage
- Full profile
Microsoft
Phi-3-vision-128k-instruct
Phi multimodal
Microsoft's Phi-3 vision model with 128K context for lightweight image understanding on edge devices.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-3.5-mini-instruct
Phi
Microsoft's Phi-3.5 models with 128K context, including a MoE variant for improved efficiency.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-3.5-MoE-instruct
Phi
Microsoft's Phi-3.5 models with 128K context, including a MoE variant for improved efficiency.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-3.5-vision-instruct
Phi multimodal
Microsoft's Phi-3.5 vision model with 128K context for image understanding and multimodal chat.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-4-mini-flash-reasoning
Phi
Microsoft's Phi-4 reasoning variants with 128K context for compact, efficient reasoning on constrained infrastructure.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-4-mini-instruct
Phi
Microsoft's Phi-4 reasoning variants with 128K context for compact, efficient reasoning on constrained infrastructure.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-4-multimodal-instruct
Phi multimodal
Microsoft's 5.6B Phi-4 multimodal model with vision, audio, and text input for lightweight assistant features.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-4-reasoning
Phi
Microsoft's Phi-4 reasoning variants with 128K context for compact, efficient reasoning on constrained infrastructure.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-4-reasoning-plus
Phi
Microsoft's Phi-4 reasoning variants with 128K context for compact, efficient reasoning on constrained infrastructure.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-4-reasoning-vision-15B
Phi multimodal
Microsoft's Phi-4 reasoning vision model (15B) combining visual understanding with chain-of-thought reasoning.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen2.5-1.5B-Instruct
Qwen2.5
Alibaba's Qwen2.5 generalist models with 131K context for multilingual apps, coding assistants, and self-hosted deployment.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen2.5-14B-Instruct
Qwen2.5
Alibaba's Qwen2.5 generalist models with 131K context for multilingual apps, coding assistants, and self-hosted deployment.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen2.5-32B-Instruct
Qwen2.5
Alibaba's Qwen2.5 generalist models with 131K context for multilingual apps, coding assistants, and self-hosted deployment.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen2.5-3B-Instruct
Qwen2.5
Alibaba's Qwen2.5 generalist models with 131K context for multilingual apps, coding assistants, and self-hosted deployment.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen2.5-72B-Instruct
Qwen2.5
Alibaba's Qwen2.5 generalist models with 131K context for multilingual apps, coding assistants, and self-hosted deployment.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen2.5-7B-Instruct
Qwen2.5
Alibaba's Qwen2.5 generalist models with 131K context for multilingual apps, coding assistants, and self-hosted deployment.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen2.5-Max
Qwen2.5
Alibaba's Qwen2.5 generalist models with 131K context for multilingual apps, coding assistants, and self-hosted deployment.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen2.5-VL-72B-Instruct
Qwen VL
Alibaba's flagship 72B vision-language model with 131K context for document reading, image understanding, and multimodal reasoning.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen2.5-VL-7B-Instruct
Qwen VL
Alibaba's efficient 7B vision-language model for lightweight multimodal assistants and document understanding.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen3-Coder-Next
Qwen3
Alibaba's Qwen3 family with thinking/non-thinking mode switching, 100+ language support, and 131K context via YaRN.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen3.5-0.8B
Qwen3
Alibaba's Qwen3 family with thinking/non-thinking mode switching, 100+ language support, and 131K context via YaRN.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen3.5-122B-A10B
Qwen3
Alibaba's Qwen3 family with thinking/non-thinking mode switching, 100+ language support, and 131K context via YaRN.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen3.5-27B
Qwen3
Alibaba's Qwen3 family with thinking/non-thinking mode switching, 100+ language support, and 131K context via YaRN.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen3.5-2B
Qwen3
Alibaba's Qwen3 family with thinking/non-thinking mode switching, 100+ language support, and 131K context via YaRN.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen3.5-35B-A3B
Qwen3
Alibaba's Qwen3 family with thinking/non-thinking mode switching, 100+ language support, and 131K context via YaRN.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen3.5-397B-A17B
Qwen3
Alibaba's Qwen3 family with thinking/non-thinking mode switching, 100+ language support, and 131K context via YaRN.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen3.5-4B
Qwen3
Alibaba's Qwen3 family with thinking/non-thinking mode switching, 100+ language support, and 131K context via YaRN.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Alibaba Qwen
Qwen3.5-9B
Qwen3
Alibaba's Qwen3 family with thinking/non-thinking mode switching, 100+ language support, and 131K context via YaRN.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Perplexity
Sonar
Sonar
Perplexity's fast, cost-efficient Sonar model for basic web search and Q&A. $1/$1 per MTok plus request fee.
- Context
- 128,000
- Input
- $0.001/1K tok
- Output
- $0.001/1K tok
- Coverage
- Full profile
Morph
warpgrep-v2
Morph WarpGrep
Morph's WarpGrep v2 for ultra-fast agentic code search. #1 on SWE-Bench Pro with sub-6s searches. 15.8% cheaper and 22% faster than alternatives.
- Context
- 128,000
- Input
- $0.0005/1K tok
- Output
- $0.0015/1K tok
- Coverage
- Full profile
DeepSeek
DeepSeek-OCR
DeepSeek multimodal
DeepSeek's specialized OCR models for document parsing, table extraction, and image-to-text conversion.
- Context
- 16,384
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-OCR-2
DeepSeek multimodal
DeepSeek's specialized OCR models for document parsing, table extraction, and image-to-text conversion.
- Context
- 16,384
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-VL2-Small
DeepSeek multimodal
DeepSeek's lightweight vision-language model for efficient multimodal understanding.
- Context
- 16,384
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
MiniMax
image-01
MiniMax media
MiniMax's image generation models for text-to-image creation, including a live animation variant.
- Context
- 8,192
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
MiniMax
image-01-live
MiniMax media
MiniMax's image generation models for text-to-image creation, including a live animation variant.
- Context
- 8,192
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
DeepSeek
Janus-Pro-7B
DeepSeek multimodal
DeepSeek's multimodal understanding and generation model (7B) for image-to-text and text-to-image tasks.
- Context
- 16,384
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Llama 3.1 405B Instruct
Llama
Meta's Llama 3.1 405B flagship with 128K context.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Llama 3.1 70B Instruct
Llama
Meta's Llama 3.1 70B with 128K context for instruction-following.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
MiniMax
MiniMax-Speech-02
MiniMax media
MiniMax's TTS model for high-fidelity speech synthesis with voice cloning capabilities.
- Context
- 8,192
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
MiniMax
music-2.0
MiniMax media
MiniMax's music generation model for AI-composed audio tracks.
- Context
- 8,192
- Input
- $0.001/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
Microsoft
Phi-4
Phi
Microsoft's 14B parameter Phi-4, a state-of-the-art small model trained on 9.8T tokens with strong reasoning on MMLU (84.8) and GPQA (56.1).
- Context
- 16,384
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
StepFun
Step-3.5-Flash
Step
StepFun's 199B parameter Step-3.5-Flash for text generation and reasoning. 91.9K downloads on HuggingFace. Available in BF16, FP8, and GGUF quantizations.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Anthropic
Claude Opus 3
Claude 200K
Anthropic's legacy Opus 3 model with 200K context.
- Context
- 200,000
- Input
- $0.02/1K tok
- Output
- $0.08/1K tok
- Coverage
- Full profile
Anthropic
Claude Opus 4
Claude 200K
Anthropic's Opus 4 with 200K context for complex reasoning and agentic tasks.
- Context
- 200,000
- Input
- $0.02/1K tok
- Output
- $0.08/1K tok
- Coverage
- Full profile
Meta
Llama 3.2 90B Vision Instruct
Llama multimodal
Meta's Llama 3.2 90B vision model with 128K context for image understanding.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
phi-1
Phi
Microsoft's earlier Phi models with shorter context windows for edge and local deployment.
- Context
- 4,096
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
phi-1_5
Phi
Microsoft's earlier Phi models with shorter context windows for edge and local deployment.
- Context
- 4,096
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
phi-2
Phi
Microsoft's earlier Phi models with shorter context windows for edge and local deployment.
- Context
- 4,096
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-3-medium-4k-instruct
Phi
Microsoft's earlier Phi models with shorter context windows for edge and local deployment.
- Context
- 4,096
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-3-mini-4k-instruct
Phi
Microsoft's earlier Phi models with shorter context windows for edge and local deployment.
- Context
- 4,096
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Microsoft
Phi-tiny-MoE-instruct
Phi
Microsoft's earlier Phi models with shorter context windows for edge and local deployment.
- Context
- 4,096
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Morph
flash-compact
Morph Flash Compact
Morph's Flash Compact for near-lossless context compaction at 33,000+ tok/sec. Reduces context 50-70% in under 2 seconds. +0.6% on SWE-Bench Pro.
- Context
- 200,000
- Input
- $0.0003/1K tok
- Output
- $0.001/1K tok
- Coverage
- Full profile
OpenAI
GPT Image 1
OpenAI image
OpenAI's natively multimodal image generation model.
- Context
- 32,768
- Input
- $0.004/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Amazon Web Services
Nova Micro
Nova
Amazon's Nova Micro model on Bedrock for ultra-fast, ultra-low-cost text inference.
- Context
- 128,000
- Input
- $0.00/1K tok
- Output
- $0.0001/1K tok
- Coverage
- Full profile
StepFun
Step3-VL-10B
Step
StepFun's 10B Step3-VL vision-language model with 211K downloads on HuggingFace. Supports FP8 quantization.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
OpenAI
gpt-audio-mini
OpenAI audio
OpenAI's mini audio models for cost-efficient voice interfaces.
- Context
- 128,000
- Input
- $0.006/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
OpenAI
gpt-realtime-mini
OpenAI audio
OpenAI's mini audio models for cost-efficient voice interfaces.
- Context
- 128,000
- Input
- $0.006/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Xiaomi
MiMo-VL-7B
MiMo
Xiaomi's MiMo-VL 7B vision-language model with 136K downloads on HuggingFace for image understanding and multimodal reasoning.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
OpenAI
chatgpt-image-latest
OpenAI image
OpenAI's lightweight image generation variants.
- Context
- 32,768
- Input
- $0.004/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
OpenAI
gpt-image-1-mini
OpenAI image
OpenAI's lightweight image generation variants.
- Context
- 32,768
- Input
- $0.004/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
OpenAI
gpt-oss-20b
gpt-oss
OpenAI's 20B open-weight model for efficient frontier reasoning on consumer hardware.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Llama 3.1 8B Instruct
Llama
Meta's Llama 3.1 8B with 128K context for efficient deployment.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Llama 3.2 11B Vision Instruct
Llama multimodal
Meta's Llama 3.2 11B vision model for efficient multimodal understanding.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Llama 3.2 1B Instruct
Llama
Meta's lightweight Llama 3.2 models for edge and mobile deployment with 128K context.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Llama 3.2 3B Instruct
Llama
Meta's lightweight Llama 3.2 models for edge and mobile deployment with 128K context.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Xiaomi
MiMo-Audio-7B
MiMo
Xiaomi's MiMo-Audio 7B for any-to-any audio and text generation. Includes base and instruct variants.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Code Llama 70B Instruct
Llama
Meta's original Llama 3 and Code Llama 70B models with 8K context.
- Context
- 8,192
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Liquid AI
LFM2-24B-A2B
LFM2
Liquid AI's flagship 24B/2.3B-active MoE hybrid model. 112 tok/s on AMD CPU, 293 tok/s on H100. Fits in 32GB RAM. Trained on 17T tokens with 30 conv + 10 attn layers. Supports 9 languages.
- Context
- 32,768
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Llama Guard 4 12B
Llama Guard
Meta's Llama Guard 4 (12B) safety model for filtering unsafe prompts.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Meta Llama 3 70B Instruct
Llama
Meta's original Llama 3 and Code Llama 70B models with 8K context.
- Context
- 8,192
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
StepFun
Step-Audio-R1.1
Step
StepFun's 33B Step-Audio-R1.1 for audio-text-to-text generation and understanding.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
OpenAI
GPT-4o mini Transcribe
OpenAI audio
OpenAI's GPT-4o based TTS and transcription models.
- Context
- 128,000
- Input
- $0.006/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
OpenAI
GPT-4o mini TTS
OpenAI audio
OpenAI's GPT-4o based TTS and transcription models.
- Context
- 128,000
- Input
- $0.006/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
OpenAI
GPT-4o Transcribe
OpenAI audio
OpenAI's GPT-4o based TTS and transcription models.
- Context
- 128,000
- Input
- $0.006/1K tok
- Output
- $0.02/1K tok
- Coverage
- Full profile
Meta
Code Llama 34B Instruct
Llama
Meta's smaller Llama 3 and Code Llama models with 8K context.
- Context
- 8,192
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Llama Guard 3 11B Vision
Llama Guard
Meta's Llama Guard 3 vision safety model for multimodal content moderation.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Meta Llama 3 8B Instruct
Llama
Meta's smaller Llama 3 and Code Llama models with 8K context.
- Context
- 8,192
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Liquid AI
LFM2-8B-A1B
LFM2
Liquid AI's 8.3B/1.5B-active MoE hybrid model with 12T training tokens. 24 layers (18 conv + 6 attn). 47.9K downloads on HuggingFace.
- Context
- 32,768
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Black Forest Labs
FLUX 1.1 Pro
FLUX
Black Forest Labs' flagship FLUX 1.1 Pro for high-quality text-to-image generation via API.
- Context
- 512
- Input
- $0.004/1K tok
- Output
- $0.004/1K tok
- Coverage
- Full profile
Black Forest Labs
FLUX 1.1 Pro Ultra
FLUX
Black Forest Labs' FLUX 1.1 Pro Ultra for ultra-high-resolution 4MP image generation via API.
- Context
- 512
- Input
- $0.006/1K tok
- Output
- $0.006/1K tok
- Coverage
- Full profile
Black Forest Labs
FLUX 1 Pro
FLUX
Black Forest Labs' original FLUX 1 Pro for high-quality text-to-image via API.
- Context
- 512
- Input
- $0.005/1K tok
- Output
- $0.005/1K tok
- Coverage
- Full profile
Liquid AI
LFM2.5-1.2B-Instruct
LFM2.5
Liquid AI's 1.2B LFM2.5-Instruct with 262K downloads on HuggingFace. Extended 128K context for edge reasoning and instruction following.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Liquid AI
LFM2.5-1.2B-Thinking
LFM2.5
Liquid AI's 1.2B LFM2.5-Thinking with chain-of-thought reasoning for edge devices. 30K downloads on HuggingFace.
- Context
- 131,072
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Liquid AI
LFM2-2.6B
LFM2
Liquid AI's 3B parameter LFM2 for lightweight text generation and tool use on edge devices.
- Context
- 32,768
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Perplexity
pplx-embed-v1-4b
Sonar embedding
Perplexity's 4B embedding model with 2560 dimensions for high-quality semantic search and RAG. $0.03 per MTok.
- Context
- 8,192
- Input
- $0.00/1K tok
- Output
- $0.00/1K tok
- Coverage
- Full profile
Perplexity
pplx-embed-v1-0.6b
Sonar embedding
Perplexity's lightweight 0.6B embedding model with 1024 dimensions for fast, cost-efficient embeddings. $0.004 per MTok.
- Context
- 8,192
- Input
- $0.00/1K tok
- Output
- $0.00/1K tok
- Coverage
- Full profile
StepFun
NextStep-1.1
Step
StepFun's 15B NextStep-1.1 for text-to-image generation. Supports image editing and modification based on text prompts.
- Context
- 512
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
Meta
Prompt Guard 86M
Llama Guard
Meta's lightweight 86M Prompt Guard for detecting jailbreak and prompt injection attempts.
- Context
- 512
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
NVIDIA
Nemotron 3 Super 120B
Nemotron
NVIDIA's flagship 120B/12B-active LatentMoE model with 1M context, trained on 25T tokens. Strong on agentic workflows, reasoning, and long-context tasks. Requires 8x H100-80GB.
- Context
- 1,048,576
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
NVIDIA
Llama Nemotron Super 49B
Llama Nemotron
NVIDIA's Llama-based Nemotron Super 49B for high-accuracy reasoning, agentic tasks, and RAG workflows.
- Context
- 131,072
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
NVIDIA
Nemotron-Cascade 2
Nemotron
NVIDIA's 32B (30B-A3B MoE) Nemotron-Cascade 2 trained with cascade RL and multi-domain on-policy distillation. 74.8K downloads on HuggingFace.
- Context
- 131,072
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Xiaomi
MiMo-V2-Flash
MiMo
Xiaomi's MiMo-V2-Flash: 309B total/15B active MoE with hybrid sliding window attention, Multi-Token Prediction, and 256K context. Scores 94.1 on AIME 2025, 73.4 on SWE-Bench. Trained on 27T tokens with 6x KV-cache reduction.
- Context
- 262,144
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
IBM
Granite 3.1 8B Instruct
Granite
IBM's Granite 3.1 8B model for enterprise text and code tasks with Apache 2.0 license.
- Context
- 128,000
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
IBM
Granite 3.0 8B Instruct
Granite
IBM's Granite 3.0 8B model for enterprise text and code generation.
- Context
- 128,000
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
NVIDIA
Llama Nemotron Nano 4B
Llama Nemotron
NVIDIA's compact 4B Llama Nemotron Nano for edge AI with high-accuracy reasoning. Runs on consumer GPUs.
- Context
- 131,072
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
NVIDIA
Nemotron 3 Nano 4B
Nemotron
NVIDIA's compact 4B Nemotron Nano for efficient local AI with hybrid Mamba-2 architecture. Runs on consumer GPUs.
- Context
- 131,072
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
IBM
Granite 3.1 2B Instruct
Granite
IBM's Granite 3.1 2B model for lightweight enterprise deployment.
- Context
- 128,000
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
NVIDIA
Llama-Embed-Nemotron 8B
Nemotron embedding
NVIDIA's Llama-Embed-Nemotron 8B ranked #1 on multilingual MTEB leaderboard with text and image retrieval support.
- Context
- 32,768
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 2 Klein 9B
FLUX.2
Black Forest Labs' 9B FLUX.2 Klein for sub-second image generation with 4 inference steps. Matches models 5x its size in quality. Requires ~29GB VRAM (RTX 4090+).
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 2 Klein 9B KV
FLUX.2
Black Forest Labs' FLUX.2 Klein 9B KV variant with key-value cache optimization for even faster inference.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 2 Klein 4B
FLUX.2
Black Forest Labs' 4B FLUX.2 Klein with Apache 2.0 license for fast, commercial-use image generation. 268K monthly downloads.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 2 Dev
FLUX.2
Black Forest Labs' 32B parameter FLUX.2 Dev, the state-of-the-art open text-to-image model with single and multi-reference editing. 987K monthly downloads on HuggingFace.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 1 Schnell
FLUX
Black Forest Labs' 12B FLUX 1 Schnell for ultra-fast image generation in 1-4 steps with Apache 2.0 license.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 1 Dev
FLUX
Black Forest Labs' 12B parameter FLUX 1 Dev for open-weight text-to-image generation with guidance distillation.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 1 Kontext Dev
FLUX.1
Black Forest Labs' FLUX 1 Kontext Dev for context-aware image editing with 95.6K downloads on HuggingFace.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
NVIDIA
NV-Embed v2
Nemotron embedding
NVIDIA's state-of-the-art text embedding model ranked #1 on MTEB leaderboard for retrieval and semantic similarity tasks.
- Context
- 32,768
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 1 Fill Dev
FLUX
Black Forest Labs' FLUX 1 Fill Dev for inpainting and outpainting with open weights.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 1 Krea Dev
FLUX
Black Forest Labs' FLUX 1 Krea Dev developed in partnership with Krea AI for enhanced creative image generation.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Stability AI
Stable Diffusion 3.5 Large
Stable Diffusion
Stability AI's flagship image generation model with high-quality text-to-image capabilities.
- Context
- 77
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 1 Redux Dev
FLUX
Black Forest Labs' FLUX 1 Redux Dev for image variation and style transfer with open weights.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 1 Canny Dev
FLUX
Black Forest Labs' FLUX 1 Canny Dev for edge-aware image generation using Canny edge detection.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Black Forest Labs
FLUX 1 Depth Dev
FLUX
Black Forest Labs' FLUX 1 Depth Dev for depth-aware image generation and structural editing.
- Context
- 512
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29
Stability AI
Stable Diffusion 3.5 Medium
Stable Diffusion
Stability AI's efficient medium-sized image generation model.
- Context
- 77
- Input
- Unpublished
- Output
- Unpublished
- Coverage
- Verified 2026-03-29