Provider analysis
DeepSeek
Lab publishing DeepSeek reasoning, coding, OCR, multimodal, and open model releases.
This provider page blends full-profile entries with broader verified listings. Use it to separate deeply evaluated flagship models from source-backed records that are tracked primarily for market visibility, access data, and freshness coverage.
Tracked models available through provider-managed APIs.
Models with downloadable weights or self-hosted distribution paths.
Total source references attached across this provider catalog.
DeepSeek
DeepSeek-V3.2
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V3.2-Exp
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V3.1
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V3.1-Base
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V3
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-V2.5
DeepSeek
DeepSeek's MoE flagship models (671B total, 37B activated params) with 128K context, strong on math, code, and reasoning benchmarks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-R1
DeepSeek R1
DeepSeek's reasoning model (671B total, 37B activated) trained with large-scale RL, achieving o1-level performance on math, code, and reasoning tasks.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-R1-Distill-Llama-70B
DeepSeek distill
DeepSeek-R1 distilled into Llama 3.3 70B, achieving strong reasoning with dense model efficiency.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-Coder-V2
DeepSeek Coder/Math
DeepSeek's specialist models for code generation and mathematical reasoning.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-Math-V2
DeepSeek Coder/Math
DeepSeek's specialist models for code generation and mathematical reasoning.
- Context
- 128,000
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
Janus-Pro-7B
DeepSeek multimodal
DeepSeek's multimodal understanding and generation model (7B) for image-to-text and text-to-image tasks.
- Context
- 16,384
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-VL2-Small
DeepSeek multimodal
DeepSeek's lightweight vision-language model for efficient multimodal understanding.
- Context
- 16,384
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-OCR
DeepSeek multimodal
DeepSeek's specialized OCR models for document parsing, table extraction, and image-to-text conversion.
- Context
- 16,384
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile
DeepSeek
DeepSeek-OCR-2
DeepSeek multimodal
DeepSeek's specialized OCR models for document parsing, table extraction, and image-to-text conversion.
- Context
- 16,384
- Input
- Not applicable
- Output
- Not applicable
- Coverage
- Full profile