DAHO
AIMarch 20, 20265 min

The Model War: GPT-5.4, Claude Opus 4.6, and Gemini 3.1

OpenAI, Anthropic and Google are shipping updates every 2-3 weeks. The capability gap has closed — choosing a model is now a workflow decision, not a capability one.

#llm#openai#anthropic#google

The race that never stops

If you were following the LLM world closely in 2024, 2026 has you sprinting. The three major labs — OpenAI, Anthropic and Google — are in a release cycle measured in weeks, not months. Every update brings new benchmarks, new capabilities and, above all, new reasons to re-evaluate your AI stack.

Here's what's on the table right now.

The three models and what makes them different

GPT-5.4 "Thinking" (OpenAI — released March 5)

OpenAI's latest model arrives with a 1 million token context and reasoning capabilities the company describes as "GPT-6 level" on certain tasks. The "Thinking" mode activates an extended reasoning chain that significantly improves results on math problems, logic and complex coding.

Most interesting: OpenAI reports emergent behavior on multimodal tasks that weren't in the original training set. The 1M token context makes it especially useful for analyzing full codebases or extensive documentation.

Claude Opus 4.6 (Anthropic — the model I use daily)

Anthropic has established itself as the preferred choice for serious developers, and Opus 4.6 reinforces that position. Context is also 1 million tokens, but where Anthropic shines is coding capabilities: following complex instructions, generating code with fewer logic errors and a lower tendency toward "hallucinations" in software architecture contexts.

Pricing is $3 per million input tokens and $15 per million output — still the premium segment benchmark.

Gemini 3.1 (Google — the benchmark leader)

Gemini 3.1's numbers are impressive: it leads 13 of the 16 main industry benchmarks and reached 77.1% on ARC-AGI-2 — the hardest test that exists for measuring general AI reasoning. That score surpasses the human baseline on that test.

Google also has the advantage of native integration with its ecosystem: Workspace, Cloud, Android. For teams already on that stack, the value proposition is clear.

What this means for developers

Last year it still made sense to say "use GPT-4 for coding, use Claude for writing." That kind of capability-based differentiation is disappearing. All three models are competent at practically everything.

What now defines the choice is workflow fit:

  • Is your team on Google Cloud and using Workspace? Gemini 3.1 has integration advantages that are hard to ignore.
  • Are you building a coding assistant or an agent that will write and review extensive code? Claude Opus 4.6 remains the most solid choice.
  • Do you need extended reasoning on complex mathematical or logical tasks? GPT-5.4 Thinking is hard to beat.

Iteration speed also matters. With updates every 2-3 weeks, any model's competitive advantage can be short-lived. The teams that win are those with the infrastructure to switch models quickly — not those who bet everything on a single provider.

How I actually use them in practice

This is what my own workflow looks like right now, as a developer who builds web and mobile products:

Day-to-day coding: Claude Opus 4.6. It follows complex multi-file instructions better than anyone else, hallucinates less on architecture decisions, and its explanations are clearer when I need to understand something I didn't write.

Research and writing: Also Claude, primarily. The reasoning is transparent — it shows you its work in a way that helps you catch errors.

Long document analysis: GPT-5.4 for the 1M token context. When I need to analyze an entire codebase or a massive API specification, nothing comes close right now.

Experimenting: I keep Gemini 3.1 open through Google AI Studio. Whenever I'm trying something at the frontier — complex reasoning, multimodal tasks — I test it there. The benchmark numbers are real and it's worth staying close.

The hidden cost: context switching tax

Here's something I've noticed that nobody talks about: switching models constantly has a hidden productivity cost. Every model has its own quirks, its response patterns, the way it interprets ambiguous instructions.

When I switch from Claude to GPT in the middle of a complex task, I spend 10-15 minutes recalibrating — rephrasing prompts, adjusting expectations, rediscovering what this model does or doesn't do well.

My recommendation: pick a primary model and stick to it for at least a month. Use others for specific tasks where they clearly shine, but avoid the trap of testing every new release just because it topped a benchmark.

Benchmarks test models on standardized tasks. Your work isn't standardized.

What's coming that nobody is talking about

Context windows are approaching the point where a model can load your entire codebase, your documentation, your past conversations, and your specifications simultaneously. That's not a future scenario — GPT-5.4 and Claude Opus 4.6 can already do it within certain limits.

When context is effectively unlimited, the bottleneck shifts from "how much can the model remember" to "how well can the model reason over that information." And that's where we're going to see the real differentiation over the next 12-18 months.

The labs that figure out reliable long-context reasoning — not just loading tokens but actually using them well — will win this phase of the race.

A final thought

We're at the strangest and most exciting moment in the history of software. The models available today would have seemed like science fiction two years ago. And in two weeks there will probably be something new.

The strategy isn't to find the best model. The strategy is to build the workflow flexibility to switch when it matters — and the discipline to not switch when it doesn't.

The Model War: GPT-5.4, Claude Opus 4.6, and Gemini 3.1