Anthropic doesn’t make a lot of noise when they release things. No countdown timers, no cryptic tweets from the CEO. On April 16th, they just… updated the model. Classic Anthropic.
Here’s what actually matters, based on real testing.
But Claude Opus 4.7 deserves more attention than it’s getting, because the improvements here aren’t incremental — they’re the kind of changes that genuinely affect how you work with the model. I’ve been using it since launch day, and there are a few things that immediately stood out.
What’s Actually New in Claude Opus 4.7?
Let me skip the marketing fluff and get to the stuff that matters.
High-Resolution Image Support (Finally)
This has been one of the most requested features for Claude, and Opus 4.7 delivers it. The maximum image resolution jumped from 1,568px / 1.15MP to 2,576px / 3.75MP. That’s more than triple the pixel count. If you’ve been working with diagrams, screenshots, documents, or any visual content, you’ll notice the difference immediately — Claude can now read text in images that it previously squinted at and got wrong.
For developers doing UI testing, designers getting feedback on mockups, or anyone who sends Claude photos of whiteboards — this is a big deal.
Adaptive Thinking Replaces Extended Thinking
Here’s a change that might trip up some API users. Opus 4.7 drops the old “extended thinking” mode entirely. Adaptive thinking is now the only thinking mode, and according to Anthropic’s internal benchmarks, it reliably outperforms extended thinking across all test categories.
There’s also a new “xhigh” effort setting that sits between “high” and “max.” It gives developers more control over the thinking depth vs. response time tradeoff. In practice, I’ve found xhigh to be the sweet spot for complex coding tasks — max effort takes noticeably longer without proportional quality gains, while high sometimes misses nuances on tricky problems.
Task Budgets (Beta)
This one’s for people building agentic systems, and it’s clever. Task budgets give Claude a rough token target for an entire agentic loop. The model sees a running countdown and uses it to prioritize work, allocate effort to the most important sub-tasks, and wrap things up gracefully as the budget gets consumed.

Before task budgets, agentic Claude would sometimes spiral — spending too many tokens on early steps and running out of context for the final, most critical actions. Task budgets don’t eliminate this problem entirely, but they make it significantly less common. If you’re building autonomous workflows, this feature alone justifies testing the upgrade.
Key Improvements: Opus 4.6 → 4.7
📸
Hi-Res Vision
3.75MP (was 1.15MP)
3.3x more pixels
🧠
Adaptive Thinking
New xhigh effort level
Beats extended thinking
📊
Task Budgets
Token-aware agentic loops
Beta feature
🏆
SWE-bench
87.6% (was 80.8%)
+8.4% improvement
SWE-bench Verified
4.6: 80.8%
4.7: 87.6%
SWE-bench Pro
4.6: 53.4%
4.7: 64.3%
GDPVal-AA Elo
Opus 4.7: 1,753
vs GPT-5.4: 1,674

Benchmark Numbers: Is It Actually Better?
Numbers don’t lie, and the benchmark improvements are significant. SWE-bench Verified — the industry’s go-to test for real-world software engineering — went from 80.8% on Opus 4.6 to 87.6% on Opus 4.7. That’s not a marginal bump; an almost 7-point improvement on a benchmark where most models fight for tenths of a percentage point is remarkable.
SWE-bench Pro tells a similar story, jumping from 53.4% to 64.3%. And on GDPVal-AA, an Elo-based knowledge work benchmark, Opus 4.7 scores 1,753 versus GPT-5.4’s 1,674. That puts Claude comfortably ahead of OpenAI’s flagship on knowledge-intensive tasks.
In my own testing — which mostly involves coding, document analysis, and research tasks — Opus 4.7 feels noticeably more capable at complex multi-step reasoning. It catches edge cases that 4.6 would miss, and it’s better at maintaining coherence over long conversations.
The Breaking Change You Need to Know About
If you’re using the Claude API, pay attention to this one. Starting with Opus 4.7, setting temperature, top_p, or top_k to any non-default value returns a 400 error. That’s right — if your code explicitly sets temperature=0.7 or something similar, it will break.
The fix is simple: just remove those parameters from your API calls entirely. Anthropic is pushing developers toward adaptive thinking, which handles temperature-like behavior internally. But if you’re running production systems with hardcoded temperature settings, test before you upgrade your model string.
Pricing: Same as Before
Good news here. Opus 4.7 launches at the same price as Opus 4.6: $5 per million input tokens and $25 per million output tokens. The 1M token context window and 128K max output tokens are unchanged. So you’re getting a meaningfully better model at zero additional cost — which is exactly what you want to see from an AI provider.
How Does Opus 4.7 Compare to GPT-5.4?
This is the matchup everyone’s watching. On pure coding benchmarks, Opus 4.7 currently leads. On creative writing and conversational tasks, it’s more subjective — both models are excellent, and preference often comes down to personal taste. GPT-5.4 still has an edge in multimodal capabilities beyond images (audio, video understanding), but Claude’s vision improvements narrow the gap significantly.
The honest answer? Both models are incredibly capable in April 2026. If you’re already in the Claude ecosystem (using Claude Code, Claude for Enterprise, etc.), upgrading to 4.7 is a no-brainer. If you’re choosing between providers, try both with your specific use case — benchmarks tell one story, but your workload tells the real one.
Should You Upgrade?
If you’re a casual Claude user on the web or mobile apps, you don’t need to do anything — Anthropic has already switched the default Opus model to 4.7. You’re already using it.
If you’re an API user, the upgrade is worth it for the benchmark improvements alone, but test your integration first because of the temperature/top_p breaking change. And if you’re building agentic systems, task budgets are reason enough to migrate.
The only people who might want to hold off are those with extensively tuned prompts that rely on specific temperature settings. In that case, rework your prompts to drop those parameters, test in a staging environment, and then migrate.
Frequently Asked Questions
When was Claude Opus 4.7 released?
Claude Opus 4.7 was released on April 16, 2026. It replaces Opus 4.6 as the default Opus model across all Claude products and API endpoints.
Does Claude Opus 4.7 cost more than 4.6?
No. Opus 4.7 is priced identically to 4.6 at $5 per million input tokens and $25 per million output tokens. Context window (1M tokens) and max output (128K tokens) are also unchanged.
What is the task budgets feature in Claude 4.7?
Task budgets is a beta feature that gives Claude a token budget for an entire agentic workflow. The model sees a running countdown and prioritizes sub-tasks accordingly, gracefully wrapping up as the budget is consumed. This prevents the model from spending too many tokens on early steps and running out of context for critical final actions.
Is Claude Opus 4.7 better than GPT-5.4?
On coding benchmarks like SWE-bench, Claude Opus 4.7 currently leads GPT-5.4. On the GDPVal-AA knowledge work benchmark, Opus 4.7 (1,753 Elo) outperforms GPT-5.4 (1,674 Elo). Creative and conversational performance is more subjective, with both models excelling in different ways.
The Bottom Line
Claude Opus 4.7 is a strong, substantive upgrade that improves on nearly every measurable dimension. The high-res vision support alone makes it significantly more useful for real-world tasks, and the SWE-bench improvements confirm that Anthropic isn’t slowing down on the technical side. If you’re already using Claude, upgrading is an easy call. If you’ve been considering switching from GPT, this is a compelling reason to run a test.
The AI model race in 2026 is tighter than ever, and Opus 4.7 keeps Claude firmly in the fight.



