Twenty-five days after Anthropic shipped Claude Opus 4.7, the community is still split on it. Reddit threads call it a regression, Notion calls it "a true teammate," and Cursor measures it at 70% on its internal CursorBench versus 58% for Opus 4.6. The truth, if you spend the model on real production app work, sits closer to the second camp than the first. This guide unpacks what Claude Opus 4.7 actually does well, where the complaints have merit, and how to drive it from Totalum's API and MCP to ship a deployable Next.js app in a single agent run.

Quick Answer
- Claude Opus 4.7, released April 16, 2026, is Anthropic's strongest agentic model. It posts a 13% lift over Opus 4.6 on Anthropic's 93-task coding benchmark and 70% on Cursor's internal CursorBench versus 58% for Opus 4.6.
- It is the model to reach for when your agent has to plan, edit, and complete a complex multi-step task without bailing out halfway. It is the wrong model for short, latency-sensitive chat.
- Pricing held steady at $5 per million input tokens and $25 per million output tokens, unchanged from Opus 4.6.
- Plugged into Totalum's MCP server, Opus 4.7 can generate, deploy, and iterate on a real production Next.js app with auth, database, file storage, and payments in one agent run.
- Use Sonnet 4.5 for routine edits and Opus 4.7 for the hard work. The cost gap is real, and routing matters more than picking a single model.
What Claude Opus 4.7 actually is
Claude Opus 4.7 is the April 2026 refresh of Anthropic's Opus family. Per the official release notes, it "carries work all the way through instead of stopping halfway." The training emphasis was multi-step planning, tool-use reliability, and long-running asynchronous agentic workflows. Visual capability also jumped: Opus 4.7 is the first Claude that natively handles images up to 2,576 pixels on the long edge, roughly 3.75 megapixels, a near-3.5x lift over Opus 4.6.
It is available wherever you would expect: direct via the Anthropic API, on Amazon Bedrock, on Google Cloud Vertex AI, on Microsoft Foundry, inside Claude Code, and inside the consumer Claude apps. The developer model ID is claude-opus-4-7. Pricing is unchanged from Opus 4.6 at $5 per million input tokens and $25 per million output tokens, so the upgrade does not penalize teams already budgeting for Opus.
What changed since Opus 4.6: the numbers that matter
Anthropic's release post leans on partner benchmarks rather than only its own. The pattern is consistent across all of them: bigger gains on long agentic runs, smaller gains on short single-turn tasks.
- Coding, Anthropic 93-task internal bench: +13% over Opus 4.6.
- CursorBench, by Cursor: 70% versus 58% for Opus 4.6.
- Rakuten-SWE-Bench production tasks: 3x more tasks resolved end-to-end than Opus 4.6.
- Databricks OfficeQA Pro document reasoning: 21% fewer errors when working with source documents.
- XBOW visual-acuity benchmark: 98.5% versus 54.5% for Opus 4.6, a near-doubling on vision-heavy agentic work.
- Notion complex multi-step workflows: +14% task success at lower token usage.
If your workload is a short prompt that needs a clever response, you will barely notice the upgrade. If your workload is "read this PDF, plan a refactor across 30 files, and submit a PR," the lift compounds across every step in the chain.
How Claude Opus 4.7 and Totalum work together
Totalum is an AI app builder that produces real Next.js codebases with auth, database, file storage, payments, and AI integrations wired in from the start. It runs as a hosted service on Cloudflare with full code ownership, custom domains, and an API plus MCP surface so any AI agent can drive it.
Two integration paths matter for Opus 4.7:
- MCP: connect Claude Code, or any MCP host, to Totalum's MCP server. Opus 4.7 then becomes the planner that drives Totalum's project, database, and deployment tools as needed. The end-to-end setup is in our Claude Code MCP tutorial.
- API: hit the Totalum HTTP API directly from a custom agent built on top of the Anthropic SDK. The agent calls Totalum endpoints to spin up a new project, push schema changes, edit pages, and trigger deployments, while Opus 4.7 plans the work.
Both paths benefit from Opus 4.7's stronger tool-use reliability. The pre-4.7 failure mode was an agent that would call a tool, get a result, then forget the goal and answer in chat instead. Opus 4.7 sticks with the plan.
A worked example: ship a production app in one Opus 4.7 run
Below is the actual prompt shape that works well with Opus 4.7 driving Totalum. Keep the goal one sentence. Keep the constraints concrete. Trust the model to plan the steps.
You are an agent that uses the Totalum MCP server to build production
web apps. Build a customer portal for a software agency that:
- Lets clients log in with email and password.
- Shows each client a list of their projects and tasks.
- Lets the client upload files and comment on tasks.
- Bills monthly via Stripe at $49 per active client.
Use the totalum-create-project tool first, then totalum-add-table,
then totalum-add-page, then totalum-deploy. Stop and ask only if a
payment secret is required from the human.
Driven from Claude Code with the Totalum MCP server registered, Opus 4.7 plans the schema, generates the pages, wires Stripe, and deploys in a single run. On internal builds we see roughly 38 minutes end-to-end for an app of this scope, with Opus 4.6 at roughly 55 minutes for the same prompt and Sonnet 4.5 at roughly 90 minutes because of tool-loop drift. The gap is not the model writing faster, it is the model giving up less.
If you want a no-MCP path, the Claude Code vs Codex comparison walks through how to wire either agent against Totalum's REST API.
Where Claude Opus 4.7 still struggles: the regression complaints, honestly
The r/ClaudeAI "Opus 4.7 is a regression" thread has 810+ comments and the complaints cluster in three patterns. Two are real, one is a routing mistake.
- Real, English-only token efficiency. A widely-shared r/ClaudeAI thread documents that prompts in German, French, and Spanish burn more tokens per useful character of output than Opus 4.6 did. Anthropic has not commented publicly. Workaround: prompt in English even if the app ships in another language.
- Real, terser refusals. Opus 4.7's safety training trips on phrasings that Opus 4.6 accepted. The fix is to be more concrete in the prompt. Ambiguous "explore X" prompts get refused, concrete "build the X table with these four columns" prompts do not.
- Routing mistake, "worse coding." Most "worse coding" posts are running Opus 4.7 against single-file refactors that Sonnet 4.5 also handles. On long agentic runs, every public benchmark says Opus 4.7 is meaningfully better. Sonnet 4.5 is the right model for short edits and was the right model before 4.7 dropped too.
Read the negative posts in aggregate and the underlying message is "I should be using Sonnet for what I was using Opus for." That is true, and is the right routing.
When to use Opus 4.7 versus Sonnet 4.5: cost and latency
Routing matters more than picking a single model. A practical rule for agent stacks driven from Totalum or Claude Code:
- Sonnet 4.5 for the routine: single-file edits, copy generation, schema-to-form scaffolding, small refactors, anything under 5 tool calls. Cost is roughly 5x lower than Opus 4.7 per token.
- Opus 4.7 for the hard: cross-file refactors, debugging mystery production bugs, plan-edit-test loops, anything that has historically failed midway with Sonnet.
- Mixed routing: planner agent on Opus 4.7, worker agents on Sonnet 4.5. This is the same pattern Cursor's Composer ships with and it cuts cost by 3-4x in our internal numbers.
For latency: Opus 4.7 is not a chatbot model. If you want a snappy reply in an end-user UI, route to Sonnet 4.5. If you want correctness on a hard problem, route to Opus 4.7 and run it asynchronously.
Claude Opus 4.7 vs Claude Opus 4.6 vs Claude Sonnet 4.5
| Dimension | Claude Opus 4.7 | Claude Opus 4.6 | Claude Sonnet 4.5 |
|---|---|---|---|
| Release date | April 16, 2026 | Late 2025 | October 2025 |
| Input price per 1M tokens | $5.00 | $5.00 | $1.00 |
| Output price per 1M tokens | $25.00 | $25.00 | $5.00 |
| Coding bench (Anthropic 93-task) | +13% vs Opus 4.6 | baseline | well below Opus 4.6 |
| CursorBench | 70% | 58% | not published by Cursor |
| Rakuten-SWE-Bench production tasks resolved | 3x Opus 4.6 | baseline | roughly equivalent to Opus 4.6 |
| Vision (XBOW visual acuity) | 98.5% | 54.5% | similar to Opus 4.6 |
| Max image resolution | 2,576 px (3.75 MP) | 1,568 px (1.1 MP) | 1,568 px |
| Multi-step task success (Notion bench) | +14% over Opus 4.6 | baseline | below Opus 4.6 |
| Best use case | long agentic runs, cross-file planning, complex multimodal work | legacy production workflows | high-volume short tasks, end-user chat |
External pricing sources: Anthropic pricing page and OpenRouter's claude-opus-4.7 page.
If you are choosing between Claude Opus 4.7 and an MCP-driven stack on Cursor, our Cursor vs Claude Code in 2026 breakdown maps the workflow differences.
FAQ
Is Claude Opus 4.7 worth the price over Sonnet 4.5?
For agentic work with more than 5-10 tool calls per task, yes. The completion-rate gap pays for the price gap quickly. For short single-step work, no, route to Sonnet 4.5.
Can I use Claude Opus 4.7 with Cursor or Claude Code?
Yes. Both Cursor and Claude Code support claude-opus-4-7 as a selectable model. If you want the model to drive a real deployable app instead of only editing files, pair it with an MCP server like Totalum's.
Is the Reddit "Opus 4.7 is a regression" consensus accurate?
Partially. The English-only token-efficiency complaint and the terser refusals are real. The "worse at coding" complaint is mostly a routing mistake: Opus 4.7 outperforms Opus 4.6 on every public agentic benchmark, but Sonnet 4.5 is the right model for the short edits people were previously using Opus for.
How does Opus 4.7 compare to GPT-5 and Codex for production app building?
GPT-5 is competitive on coding and slightly behind on long agentic runs in current public benchmarks. Codex CLI is the closest direct comparison from OpenAI. We walked through the practical tradeoffs in Claude Code vs Codex in 2026.
Where can I run Claude Opus 4.7?
Anthropic API, Amazon Bedrock, Google Cloud Vertex AI, Microsoft Foundry, Claude Code, and the Claude apps. The developer model ID is claude-opus-4-7.
Does Totalum require a specific Claude model?
No. Totalum's MCP and API are model-agnostic. You can drive Totalum from Claude Opus 4.7, Sonnet 4.5, GPT-5, Codex, Cursor's agent, or any custom agent. Opus 4.7 currently posts the highest end-to-end completion rate on our internal builds.
Ready to build with Claude Opus 4.7 and Totalum?
If you want to pair Anthropic's strongest agent model with a builder that ships production Next.js apps, the path is short. Register at totalum.app, connect your Claude Code or custom agent to the Totalum MCP server, and run an Opus 4.7 plan against a real project. The agent does the work, you keep the code.