Deepseek V4 Flash
Deepseek V4 Flash
All models
DeepSeek-V4-Flash is a preview of the DeepSeek-V4 series, a Mixture-of-Experts model with 284B total parameters and 13B activated, built for efficient reasoning across a 1M-token context window.
Deepseek V4 Flash is available through Ollama for local agent workflows, with support for text input. DeepSeek-V4-Flash is a preview of the DeepSeek-V4 series, a Mixture-of-Experts model with 284B total parameters and 13B activated, built for efficient reasoning across a 1M-token context window.
Deepseek V4 Flash is a local model entry from Ollama that Agent Mag tracks for install commands, available tags, modalities, and agent workflow fit. Builders can install it with the Agent Mag CLI and run it through Ollama on their own machine.
DeepSeek-V4-Flash is a preview of the DeepSeek-V4 series, a Mixture-of-Experts model with 284B total parameters and 13B activated, built for efficient reasoning across a 1M-token context window.
| Tag | Size | Context | Input |
|---|---|---|---|
| deepseek-v4-flash:cloud | - | 1M | text |
DeepSeek-V4-Pro is a frontier Mixture-of-Experts model with a 1M-token context window and three reasoning modes.
DeepSeek-OCR is a vision-language model that can perform token-efficient OCR.
DeepSeek-V3.2, a model that harmonizes high computational efficiency with superior reasoning and agent performance.
DeepSeek-V3.1-Terminus is a hybrid model that supports both thinking mode and non-thinking mode.
Compare pricing, local installs, context windows, and modality filters across the full model catalog.
Find frameworks, SDKs, and infrastructure tools that pair with this model in production workflows.
See Agent Mag coverage of model benchmarks, agent frameworks, and deployment patterns.