AI experts sharing free tutorials to accelerate your business.
← Back to News
Breaking

DeepSeek R2 Hits 92.7% on AIME With Just 32B Parameters

Krasa AI

2026-04-12

5 minute read

DeepSeek R2 Hits 92.7% on AIME With Just 32B Parameters

DeepSeek has released R2, a 32-billion-parameter open-weight reasoning model that scores 92.7% on AIME 2025 — one of the hardest math competition benchmarks in AI — while running on a single consumer GPU with 24 GB of VRAM. The model is available under an MIT license, and its API pricing undercuts comparable Western models by roughly 70%.

Why this matters: DeepSeek R2 proves that frontier-level mathematical reasoning doesn't require hundreds of billions of parameters or expensive enterprise hardware. It inverts one of the core assumptions of the post-GPT-4 era: that bigger is always better.

The Numbers That Matter

A score of 92.7% on AIME 2025 means R2 correctly solves roughly 14 out of 15 problems, where each problem demands multi-step symbolic reasoning that trips up most humans, let alone AI models. On MATH-500, another rigorous mathematical reasoning benchmark, R2 scores 89.4%.

For context, DeepSeek's previous reasoning model, R1, was a 671-billion-parameter Mixture-of-Experts behemoth released in January 2025. R2 achieves dramatically better results at a fraction of the size — 32 billion dense parameters versus 671 billion MoE parameters. That's not an incremental improvement. It's a fundamental rethinking of how reasoning capability scales.

How DeepSeek Did It

The key insight behind R2 is that most of the model's intelligence comes from post-training rather than raw scale. DeepSeek used a refined version of the GRPO (Group Relative Policy Optimization) reinforcement learning pipeline that they pioneered with R1, but applied it far more aggressively to a smaller, denser base model.

In simpler terms: instead of building a massive model and hoping reasoning emerges from scale, DeepSeek built a compact model and then specifically trained it to reason through a specialized reinforcement learning process. The model learned to break problems into steps, verify its own work, and backtrack when it hits dead ends — all through RL training rather than sheer parameter count.

This approach has significant implications for the industry. If post-training techniques can deliver frontier reasoning at 32B parameters, the cost structure of AI development shifts dramatically. Training smaller models is cheaper. Running them is cheaper. And deploying them on consumer hardware becomes feasible.

Running on Consumer Hardware

R2 runs at full speed on a single 24 GB consumer GPU — the kind of card you'd find in a high-end gaming PC or a modest workstation. That means individual researchers, small startups, and academic labs can run a frontier-class reasoning model without cloud computing costs.

This is a meaningful democratization of capability. Until now, the best reasoning models required either expensive API subscriptions or multi-GPU server setups costing tens of thousands of dollars. R2 brings that capability to hardware that costs under $2,000.

Why this matters: the cost barrier for frontier AI reasoning just dropped by an order of magnitude. Students, independent researchers, and bootstrap startups now have access to mathematical and logical reasoning capabilities that were previously reserved for well-funded labs and enterprise customers.

The Pricing Pressure

DeepSeek's API pricing for R2 undercuts comparable Western reasoning models by approximately 70%. For developers and companies that prefer API access over self-hosting, this creates immediate cost pressure on OpenAI, Anthropic, and Google.

The pricing strategy follows DeepSeek's established playbook. With R1, the company offered dramatically lower prices than Western competitors and captured significant market share among cost-sensitive developers. R2 extends that strategy to the next generation of reasoning models, with even better performance at even lower prices.

For enterprise customers evaluating reasoning model options, the calculus is straightforward: R2 offers comparable or superior mathematical reasoning at a fraction of the cost. The trade-offs are in ecosystem integration, safety features, and support — areas where Western providers still hold advantages.

The Geopolitical Context

R2's release comes at a sensitive moment. Just days ago, OpenAI, Anthropic, and Google announced a joint initiative through the Frontier Model Forum to combat Chinese AI companies using adversarial distillation (extracting knowledge from Western models to train cheaper alternatives). DeepSeek was specifically named as one of the companies engaged in this practice.

DeepSeek's response has been to keep shipping impressive models. Whether R2's capabilities were developed entirely independently or benefited from distillation techniques is a matter of active debate in the AI research community. What's not debatable is the result: a small, cheap, open-weight model that matches or exceeds Western frontier reasoning on key benchmarks.

What Comes Next

R2 is available now on Hugging Face under an MIT license. Developers can download the weights and run the model locally, or access it through DeepSeek's API at the discounted pricing.

For the broader industry, R2 raises a question that's becoming increasingly difficult to ignore: if frontier reasoning can be achieved at 32B parameters on consumer hardware, what exactly are we paying for with 700B+ parameter closed models? The answer might be reliability, safety guardrails, enterprise support, and ecosystem integration — but DeepSeek is forcing Western labs to articulate that value proposition explicitly.

The Bottom Line

DeepSeek R2 is a 32-billion-parameter model that reasons at the frontier level, runs on a gaming GPU, and costs a fraction of what Western alternatives charge. It's open-weight, MIT-licensed, and available right now. Whether you view it as a democratization of AI capability or a competitive threat to Western AI dominance probably depends on where you sit — but either way, it's a model worth paying attention to.

#AI#DeepSeek#Reasoning#Open Source#LLM

Related Articles