Back to Blog

The AI Model Race Is Getting Absurd (A Recap of Everything That Just Happened)

March 6, 2026Jeff Conn
AILLMsGPT-5ClaudeGeminiLlama

If you blinked at any point over the last 6 months, you missed about 47 major model releases. I'm only slightly exagerating. The pace of AI model development has gone from "fast" to "genuinely hard to keep up with" and I wanted to take a minute to break down what actually matters.

AI chip and neural network visualization

OpenAI: GPT-5.2 and Going Open Source?

Yeah, you read that right. OpenAI — the company that literally has "open" in the name but hasn't been open about anything in years — actually released open-weight models. GPT-oss-120b and GPT-oss-20b dropped under Apache 2.0, optimized for agentic workflows.

But the big headline is GPT-5.2. 400K token context window (up from 128K in GPT-4), perfect scores on AIME 2025 math benchmarks, and hallucination rates down to 6.2%. They also added six different reasoning levels you can dial up or down depending on wether you need speed or depth. Pretty clever actually.

Anthropic: Claude 4.5 Is a Coding Machine

Full disclosure — we use Claude extensively at Triple 3 Labs, so I'm biased. But the numbers speak for themselves. Claude Opus 4.5 hit 80.9% on SWE-bench, which means it can autonomously resolve 4 out of 5 real GitHub issues without human help.

Code on a dark screen

The "extended thinking mode" is genuinely cool — it lets the model do deliberate reasoning and self-reflection loops before answering. And the stat that blew my mind: Claude can run autonomous work for 30+ hours straight. We've personally used this for some pretty complex builds and it just... works.

Google: Gemini 3 Flash Is the Sleeper Hit

Everyone talks about Gemini 3 Pro, and sure, the 1M token context window and perfect AIME scores are impressive. But here's the thing nobody is talking about enough: Gemini 3 Flash beats Pro on 18 out of 20 benchmarks while being 3x faster.

That's wild. The "cheap" model outperforming the flagship? Google's clearly figured out something intresting with their efficiency game. Also Veo 3 for video generation is legitimately impressive but that's a whole other post.

Meta: Llama 4 Goes Mixture-of-Experts

Meta went full MoE (mixture-of-experts) with Llama 4, and the results are kinda nuts. Llama 4 Scout has 17B active parameters but supports a 10 million token context window. Ten. Million. That's like feeding it an entire codebase and then some.

AI generated face with neural network patterns

Maverick (the bigger sibling) beats GPT-4o and Gemini 2.0 Flash across a bunch of benchmarks at less than half the active parameters. The efficiency gains here are real and they matter for anyone running models on their own hardware.

The Dark Horses

Can't write this without mentioning a few others:

  • DeepSeek R1 — Reasoning at 27x cheaper cost. If you're doing high-volume work on a budget, this is your model.
  • Qwen3 — Alibaba's 235B parameter MoE model. One of the strongest open models out there right now.
  • Mistral 3 — 92% of GPT-5.2's performance at roughly 15% of the price. The value play.
  • Grok-4.1 — xAI is hanging right around Gemini 3 territory. It's getting harder to count them out.

What Does This Actually Mean?

We've moved from "one model to rule them all" to a genuinely multi-model world. The smart play isn't picking one — it's knowing which model to use for which task. Need cheap reasoning at scale? DeepSeek. Need rock-solid code generation? Claude. Need a massive context window? Gemini or Llama.

The agentic AI market is projected to hit $52 billion by 2030. The infrastructure layer is basicaly built. Now it's all about the applications — and that's where things get really exciting.