5 min leer

Claude Opus 4.7 Just Shipped. Here's What You Need To Know

Most model upgrades are incremental. Better benchmarks, a press release, maybe a few percentage points on a leaderboard nobody outside ML Twitter actually checks. You upgrade when you get around to it.

Claude Opus 4.7 is different. Anthropic shipped breaking API changes alongside the capability improvements, which means enterprise teams running agents in production cannot just swap the model ID and move on. The upgrade is real, but so is the migration work.

What shipped in Claude Opus 4.7

Anthropic released Claude Opus 4.7 on April 16, available across the Claude API, Amazon Bedrock, Google Cloud Vertex AI, and Microsoft Foundry. Pricing stays the same as Opus 4.6: $5 per million input tokens and $25 per million output tokens.

3x vision resolution. The model now accepts images up to 2,576 pixels on the long edge, roughly 3.75 megapixels, more than three times the previous 1,568-pixel limit. For AI agents that process documents, read screenshots, or extract data from complex diagrams, this is the single most impactful change. Coordinates now map 1:1 to actual pixels, eliminating the scale-factor math that tripped up earlier vision workflows.

Task budgets. A new beta feature that lets developers set an advisory token budget across a full agentic loop, including thinking, tool calls, results, and final output. The model sees a running countdown and self-moderates. This is not a hard cap. It is a suggestion the model is aware of. For enterprise teams managing agent fleet costs, this is the first native mechanism for controlling token spend without truncating capability.

New xhigh effort level. A finer control point between high and max reasoning. Claude Code now defaults to xhigh for all plans, signaling that Anthropic sees this as the sweet spot for agentic coding work.

Better memory. The model is measurably better at writing and using file-system-based memory across multi-session work. Agents that maintain scratchpads, notes, or structured memory stores should see improvement without any prompt changes.

More literal instruction following. Opus 4.7 interprets instructions more literally than its predecessor, particularly at lower effort levels. It will not silently generalize an instruction from one item to another or infer requests you did not make. This is a feature for precision-critical enterprise workflows and a potential gotcha for teams with loosely written prompts.

What breaks

This is where Opus 4.7 diverges from a typical model upgrade.

Extended thinking budgets are gone. Setting explicit thinking budgets now returns a 400 error. Adaptive thinking is the only supported mode, and it is off by default. Teams that relied on explicit thinking budgets need to migrate to adaptive thinking with effort levels.

Sampling parameters are gone. Setting temperature, top_p, or top_k to any non-default value returns a 400 error. If your agent pipelines set temperature to zero for determinism, that code breaks on upgrade. The migration path is to remove these parameters entirely and use prompting to guide behavior.

New tokenizer. Input token counts increase by roughly 1x to 1.35x depending on content. The same prompt costs up to 35% more tokens to process. This affects cost projections, context window budgets, and any hardcoded max_tokens values tuned for Opus 4.6.

Thinking content hidden by default. Thinking blocks still appear in the stream, but their content is empty unless you explicitly opt in. Products that stream reasoning to users will see a long pause before output begins unless they update.

What this means for enterprise agent deployments

Three implications worth thinking through.

Vision-heavy agents just got dramatically better

Document processing, invoice extraction, screenshot verification, chart analysis, compliance document review. Any agent workflow that involves reading images at detail is meaningfully improved. The 3x resolution increase is not marginal. It is the difference between an agent that can read a dense financial statement and one that cannot.

For enterprises deploying AI agents in regulated industries like banking and insurance, where document accuracy is non-negotiable, this is probably the most consequential capability shipped in any model update this year.

The breaking changes are a forcing function for platform abstraction

If you are running agents directly against the Anthropic API, the Opus 4.7 migration is real work. You need to audit every pipeline for thinking budgets, sampling parameters, tokenizer assumptions, and thinking display settings. Multiply that across dozens of agents and the upgrade becomes a project, not a toggle.

This is exactly the scenario where model-agnostic agent platforms earn their keep. When the orchestration layer handles model abstraction, a model upgrade is a configuration change, not an engineering sprint. The platform absorbs the breaking changes so the business workflows do not have to.

Task budgets signal the beginning of cost-aware agentic AI

Until now, controlling agent costs meant either capping tokens bluntly (which degrades quality) or letting the model run unbounded (which degrades budgets). Task budgets are the first native mechanism that lets the model itself understand its spending constraint. It is advisory, not enforced, but the concept is right.

For enterprise deployments running hundreds of agents, per-task cost governance is a prerequisite for scaling. Anthropic building cost-awareness into the model itself validates that it is a first-class requirement for production agent infrastructure, not an afterthought.

Benchmarks worth noting

Early testers reported:

  • 13% resolution improvement on coding benchmarks (93-task evaluation)

  • 3x more production tasks resolved than Opus 4.6 on Rakuten-SWE-Bench

  • 21% fewer document reasoning errors on Databricks' OfficeQA Pro

  • 10 to 15% task success improvement on complex multi-step workflows

The document reasoning improvement tracks with the vision upgrade. The coding improvements are notable because Opus 4.6 was already strong. Compounding gains on an already-capable model matter more than first-generation jumps.

What to do now

If you are running agents on Opus 4.6, do not upgrade blindly. Audit your pipelines for the four breaking changes: thinking budgets, sampling parameters, tokenizer, and thinking display. Test in staging before production.

If you are evaluating Claude for new agent deployments, start with Opus 4.7. The adaptive thinking model is cleaner than the old explicit-budget approach, and the vision improvements alone justify it for any document-heavy workflow.

If you are running agents on a platform that handles model abstraction, ask your vendor when they are shipping Opus 4.7 support and whether the breaking changes affect your deployments. The answer should be "already done" or "this week."

The model layer is the fastest-moving part of the enterprise AI stack. The teams that treat model upgrades as routine maintenance, rather than engineering projects, are the ones shipping agents at scale.

Empieza hoy

Empezar a crear agentes de IA para automatizar procesos

Únase a nuestra plataforma y empiece a crear agentes de IA para diversos tipos de automatizaciones.

Empieza hoy

Empezar a crear agentes de IA para automatizar procesos

Únase a nuestra plataforma y empiece a crear agentes de IA para diversos tipos de automatizaciones.