DeepSeek
6 articles tagged with DeepSeek
DeepSeek Releases V4-Flash-Base: 292B Parameter Base Model
DeepSeek has released V4-Flash-Base, a 292 billion parameter base model now available on Hugging Face. The model uses BF16, I64, F32, and F8_E4M3 tensor types and is distributed in Safetensors format.
DeepSeek releases V4 model preview with agent optimization, pricing undisclosed
DeepSeek released a preview of its V4 large language model on April 24, 2026, available in 'pro' and 'flash' versions. The Hangzhou-based company claims the open-source model achieves strong performance on agent-based tasks and has been optimized for tools like Anthropic's Claude Code and OpenClaw.
DeepSeek Releases V4-Pro-Base with 1.6 Trillion Parameters
DeepSeek has released DeepSeek-V4-Pro-Base, a 1.6 trillion parameter foundation model now available on Hugging Face. The base model uses BF16 precision for weights and includes support for F8_E4M3, I64, and F32 tensor types.
DeepSeek Releases V4 Pro: 1.6T Parameter MoE Model with 1M Token Context at $1.74/M Input Tokens
DeepSeek has released V4 Pro, a Mixture-of-Experts model with 1.6 trillion total parameters and 49 billion activated parameters. The model supports a 1-million-token context window and costs $1.74 per million input tokens and $3.48 per million output tokens.
DeepSeek V4 Flash Released: 284B Parameter MoE Model with 1M Context Window at $0.14 per Million Tokens
DeepSeek has released V4 Flash, a Mixture-of-Experts model with 284B total parameters and 13B activated parameters per request. The model supports a 1,048,576-token context window and is priced at $0.14 per million input tokens and $0.28 per million output tokens.
DeepSeek Releases V4-Flash and V4-Pro Models as Tencent Ships Hy3-Preview
DeepSeek has released two new models in its V4 series: DeepSeek-V4-Flash and DeepSeek-V4-Pro, both now available on Hugging Face. Separately, Tencent has shipped Hy3-Preview, marking simultaneous releases from two major Chinese AI labs.