DeepSeek V4 Released: Trillion-Parameter Open-Source Multimodal Model with Million-Token Context Window

DeepSeek has released V4, its first major model since January 2025, marking a leap to multimodal capabilities including text, image, and video generation. With approximately one trillion parameters and a one-million token context window, V4 was co-optimized with Huawei and Cambricon for Chinese AI chip hardware. The open-source release challenges both Western frontier models and China's domestic competitors, signaling that the DeepSeek phenomenon—delivering top-tier performance at dramatically lower costs—continues to accelerate.

DeepSeek's Multimodal Ambition

According to TechNode citing the Financial Times, DeepSeek plans to release V4—its first major update since V3 captured global attention in January 2025. V4 marks DeepSeek's critical leap from pure text to multimodal capabilities.

Technical Breakthroughs

Parameter Scale: V4 features approximately one trillion parameters (1T), up ~49% from V3's 671B. Given DeepSeek's MoE architecture, actual activated parameters are likely far fewer, keeping inference costs manageable.

Multimodal Capabilities: V4 introduces text, image, and video generation—DeepSeek's first foray into multimodal, directly competing with GPT-5.4, Gemini 3.1, and Claude 4.

Million-Token Context: The 1M token context window enables processing of ultra-long documents, complete codebases, and complex multi-turn conversations.

Domestic Chip Optimization

Most notably, V4 was co-optimized with Huawei and Cambricon for their latest AI chips. This means V4 runs efficiently on Huawei Ascend and Cambricon MLU chips, reducing Nvidia GPU dependency and providing a viable compute sovereignty path for China's AI industry under tightening US export controls.

Industry Impact

DeepSeek V3 shocked the industry with its reportedly $5.6M training cost. V4 continues this trajectory as an open-source release. For China's AI ecosystem, domestic chip optimization is a milestone. For global markets, a trillion-parameter open-source multimodal model will further compress commercial API premiums.

In-Depth Analysis and Industry Outlook

From a broader perspective, this development reflects the accelerating trend of AI technology transitioning from laboratories to industrial applications. Industry analysts widely agree that 2026 will be a pivotal year for AI commercialization. On the technical front, large model inference efficiency continues to improve while deployment costs decline, enabling more SMEs to access advanced AI capabilities. On the market front, enterprise expectations for AI investment returns are shifting from long-term strategic value to short-term quantifiable gains.

However, the rapid proliferation of AI also brings new challenges: increasing complexity of data privacy protection, growing demands for AI decision transparency, and difficulties in cross-border AI governance coordination. Regulatory authorities across multiple countries are closely monitoring these developments, attempting to balance innovation promotion with risk prevention. For investors, identifying AI companies with truly sustainable competitive advantages has become increasingly critical as the market transitions from hype to value validation.