DeepSeek unveils V4 Flash and V4 Pro models, claiming open‑weight lead

DeepSeek unveils V4 Flash and V4 Pro models, claiming open‑weight lead
TechCrunch

Key Points

  • DeepSeek released two preview models: V4 Flash (284 B parameters) and V4 Pro (1.6 T parameters).
  • Both support a 1 million‑token context window for extensive prompts.
  • Mixture‑of‑experts architecture activates only a fraction of parameters per task, lowering inference costs.
  • V4 Pro is the largest open‑weight model available, surpassing Moonshot AI’s Kimi K 2.6.
  • Benchmark tests show the V4 series narrowing the gap with leading closed‑source models on reasoning tasks.
  • Knowledge tests place V4 behind GPT‑5.4 and Gemini 3.1 Pro by roughly three to six months.
  • Pricing: V4 Flash at $0.14/$0.28 per million tokens; V4 Pro at $0.145/$3.48 per million tokens.
  • Both models handle text only, unlike many competitors that also process audio, video, and images.
  • Launch follows U.S. accusations of Chinese AI IP theft and prior claims that DeepSeek copies other labs' models.

Chinese AI lab DeepSeek released two preview versions of its next‑generation large language model, DeepSeek V4 Flash and V4 Pro. Both models use a mixture‑of‑experts architecture and support a 1‑million‑token context window, enabling users to feed entire codebases or long documents into prompts. DeepSeek says V4 Pro, with 1.6 trillion parameters (49 billion active), is the largest open‑weight model on the market, while V4 Flash offers a smaller, more affordable option. The company claims the new models narrow the performance gap with leading closed‑source systems and are priced well below competing frontier models.

DeepSeek, the Chinese artificial‑intelligence research lab, rolled out preview versions of its latest large language model family on Tuesday, introducing DeepSeek V4 Flash and DeepSeek V4 Pro. Both models employ a mixture‑of‑experts design that activates only a subset of parameters for each task, a strategy that trims inference costs without sacrificing capability.

Each model supports a massive 1 million‑token context window, a size that lets developers embed entire code repositories or lengthy documents within a single prompt. The smaller V4 Flash packs 284 billion total parameters, of which 13 billion are active at any moment. Its larger sibling, V4 Pro, boasts 1.6 trillion total parameters and 49 billion active ones, making it the biggest open‑weight model currently available. By comparison, Moonshot AI’s Kimi K 2.6 has 1.1 trillion parameters and MiniMax’s M1 sits at 456 billion.

DeepSeek says architectural tweaks give the V4 series a measurable edge over its predecessor, V3.2, which carried 671 billion parameters. The lab claims the new models have “almost closed the gap” with leading closed‑source systems on reasoning benchmarks. In head‑to‑head tests, V4‑Pro‑Max reportedly outperformed open‑source rivals across a suite of logical tasks and even edged out OpenAI’s GPT‑5.2 and Google’s Gemini 3.0 Pro on selected benchmarks. Coding competitions showed performance comparable to GPT‑5.4.

Knowledge‑based evaluations tell a slightly different story. The V4 models lag behind the latest frontier offerings—OpenAI’s GPT‑5.4 and Google’s Gemini 3.1 Pro—by an estimated three to six months in raw factual recall. DeepSeek acknowledges the gap, framing it as a “developmental trajectory” rather than a flaw.

Pricing is a standout feature. DeepSeek V4 Flash costs $0.14 per million input tokens and $0.28 per million output tokens, undercutting the cheapest tiers of GPT‑5.4 Nano, Gemini 3.1 Flash, and Claude Haiku 4.5. V4 Pro’s rates sit at $0.145 per million input tokens and $3.48 per million output tokens, still cheaper than Gemini 3.1 Pro, GPT‑5.5, and Claude Opus 4.7. Both models currently handle text only, lacking the multimodal capabilities—audio, video, image generation—found in many closed‑source competitors.

The launch arrives amid heightened scrutiny of Chinese AI firms. U.S. officials recently accused China of large‑scale intellectual‑property theft targeting American AI labs. DeepSeek itself has faced allegations from Anthropic and OpenAI of “distilling”—essentially copying—their models. Despite the controversy, the company positions the V4 series as a cost‑effective alternative for enterprises and developers seeking open‑weight solutions.

#Artificial Intelligence#Large Language Model#Open‑weight AI#DeepSeek#Mixture of Experts#Token Context Window#AI Benchmark#Model Pricing#AI Competition#Intellectual Property
Generated with  News Factory -  Source: TechCrunch

Also available in:

DeepSeek unveils V4 Flash and V4 Pro models, claiming open‑weight lead | AI News