deepseek
7 articles tagged with deepseek
DeepSeek V4 cuts inference costs with 1.6T parameter model using 13.7x less memory than V3
DeepSeek released V4 in two versions: a 284 billion parameter Flash model and a 1.6 trillion parameter Pro model with 49 billion active parameters. According to DeepSeek, the models use 9.5x-13.7x less memory than V3 through compressed attention mechanisms and FP4/FP8 mixed precision, while supporting a 1 million token context window.
DeepSeek V4 Pro launches with 1.6T parameters at $1.74/M tokens, undercutting Claude Sonnet 4.6 by 42%
DeepSeek released two preview models: V4 Pro (1.6T total parameters, 49B active) and V4 Flash (284B total, 13B active), both with 1 million token context windows. V4 Pro is priced at $1.74/M input tokens and $3.48/M output—42% cheaper than Claude Sonnet 4.6—while V4 Flash at $0.14/$0.28 per million tokens undercuts all small frontier models.
DeepSeek Releases V4-Flash: 284B-Parameter MoE Model With 1M Token Context at 27% Inference Cost
DeepSeek released two Mixture-of-Experts models: V4-Flash with 284B total parameters (13B activated) and V4-Pro with 1.6T parameters (49B activated). Both models support one million token context windows and use a hybrid attention architecture that requires only 27% of the inference FLOPs compared to DeepSeek-V3.2 at 1M token context.
DeepSeek Releases V4-Pro: 1.6T Parameter MoE Model with 1M Token Context
DeepSeek released two new Mixture-of-Experts models: DeepSeek-V4-Pro with 1.6 trillion parameters (49B activated) and DeepSeek-V4-Flash with 284B parameters (13B activated), both supporting one million token context length. The models achieve 27% of inference FLOPs and 10% of KV cache compared to DeepSeek-V3.2 at 1M context through a hybrid attention architecture combining Compressed Sparse Attention and Heavily Compressed Attention.
Deepseek v4 launching on Huawei chips exclusively, signaling China's AI independence progress
Deepseek v4 is launching in the coming weeks running exclusively on Huawei chips, marking a major milestone in China's effort to reduce dependency on foreign semiconductors. Chinese tech giants including Alibaba, Bytedance, and Tencent have ordered hundreds of thousands of Huawei Ascend 950PR units to deploy the model through their cloud services.
Continue v1.0.65 fixes JetBrains IDE freezes, adds DeepSeek Reasoner support
Continue released v1.0.65 for JetBrains IDEs and v1.3.36 for VS Code, focusing on stability fixes. The update addresses IDE freezing issues in JetBrains, adds compatibility with DeepSeek's reasoning model, and fixes multiple memory leaks and crash conditions.
DeepSeek releases R1 reasoning model with chain-of-thought capabilities
DeepSeek has released DeepSeek-R1, a text generation model featuring reasoning capabilities through chain-of-thought processing. The model was published January 20, 2025 and has accumulated over 830,000 downloads on Hugging Face.