Claude Opus 4.7 Is Here: Anthropic’s Latest Model Delivers, But It’s a Token Eating Machine

Claude Opus 4.7 Is Here: Anthropic’s Latest Model Delivers, But It’s a Token Eating Machine

Source: Decrypt

Published:17:42 UTC

BTC Price:$74380.3

AI Technology Innovation

Analysis

Price Impact

Low

This news concerns a new ai model release by anthropic, not a cryptocurrency. while advanced ai can indirectly impact various industries including those using crypto, there's no direct or immediate link to a specific cryptocurrency's price.

Trustworthiness

High

Price Direction

Neutral

There is no direct impact on any cryptocurrency prices from this news. the development of ai models is a separate technological advancement.

Time Effect

Long

While the immediate price impact on crypto is neutral, the long-term implications of advanced ai like claude opus 4.7 could lead to increased automation, new technological integrations, and potentially new use cases for decentralized technologies, which could indirectly influence the crypto market over an extended period.

Original Article:

Article Content:

In brief Anthropic just released its most capable Opus model yet, Claude Opus 4.7. The model delivers strong benchmark gains across coding and reasoning, but is not the controversial Mythos model that Anthropic offers to select partners. Claude Opus 4.7 shows visible chain-of-thought and unusually high token usage. Anthropic shipped Claude Opus 4.7 today, calling it the company’s most capable Opus model yet. We tested it, and the marketing lines up with the results. "Our latest model, Claude Opus 4.7, is now generally available." the company said in its official announcement. "Users report being able to hand off their hardest coding work—the kind that previously needed close supervision—to Opus 4.7 with confidence." The model arrives on the heels of weeks of user complaints about Opus 4.6 allegedly losing its edge. Developers across GitHub , Reddit, and X documented what they called " AI shrinkflation "—the feeling that the model they'd been paying for had quietly gotten worse. As we reported yesterday , Anthropic was already preparing 4.7 while sitting on something far more powerful that it can't release publicly: Claude Mythos.  When the announcement dropped this morning, X users who had been loudest about 4.6's degradation were quick to reply with sarcasm: Opus 4.7, some joked, felt like "early Opus 4.6"—the version people actually liked, before they believed Anthropic quietly turned the dials down. Anthropic, of course, has denied ever degrading model weights to manage compute demand. Welcome back opus 4.6 pic.twitter.com/hpwNkrq1tD — Dev Ed (@developedbyed) April 16, 2026 Benchmarks back up Anthropic's claims. On SWE-bench Multilingual, a benchmark that measures coding skills, Opus 4.7 scored 80.5% against 4.6's 77.8%. On GDPVal-AA, a third-party evaluation of economically valuable knowledge work across finance and legal domains, 4.7 scored 1,753 Elo against GPT-5.4's 1,674—a clear margin over the closest competitor. Document reasoning via OfficeQA Pro showed the starkest jump: 80.6% for 4.7 versus 57.1% for 4.6, with GPT-5.4 and Gemini 3.1 Pro trailing at 51.1% and 42.9% respectively. Long-term coherence on Vending-Bench 2, a benchmark that measures how good models are at long context and reasoning tasks like owning a vending business, clocked in at $10,937 money balance versus $8,018 for 4.6—a proxy for how well the model sustains useful behavior over long autonomous runs. Cybersecurity is the one area where Anthropic deliberately held back. Opus 4.7 launches with automated safeguards that detect and block prohibited or high-risk cybersecurity requests. Anthropic confirmed it "experimented with efforts to differentially reduce" 4.7's cyber capabilities during training. Security professionals can apply to a new Cyber Verification Program for access to those features. This is the company's test run for the safeguards it will eventually need to deploy with Mythos-class models at scale. Opus 4.7 is the most powerful model publicly available. Mythos Preview, Anthropic's true frontier model, remains restricted to vetted security firms. As the UK's AI Security Institute evaluated last week , Mythos was the first AI to complete "The Last Ones," a 32-step corporate network attack simulation that typically takes human red teams 20 hours. Opus 4.7 is not that. But it's the public-facing model that Anthropic will use to learn how those safety guardrails hold up in the wild before it dares release anything scarier. On the token side, Opus 4.7 uses an updated tokenizer that can map the same input to roughly 1.0x–1.35x more tokens depending on content type. The model also reasons more at higher effort levels, particularly on later turns in agentic workflows. Anthropic published a migration guide for developers planning to upgrade from 4.6. We ran our own test—the same game-building prompt we've used to evaluate every major model release. Opus 4.7 produced the best result we've ever gotten from any model. The most visually polished game, the most genuinely challenging difficulty curve, the best mechanics, and the most creative win and loss screens. It appeared to generate levels procedurally, and none of them felt impossible—a balance that has tripped up other models repeatedly. You can test the game here Emerge: The Game, created by Claude Opus 4.7 It wasn't zero-shot. Opus 4.6 had cleared that same test without any fixes. Opus 4.7 needed one round of bug fixes. That could be bad luck—a single iteration is a thin sample—but it's worth noting. What struck us more was how the model handled that round: It spotted additional bugs on its own, without being guided toward them. Opus 4.6 typically waited to be told where to look. Xiaomi MiMo v2 Pro was the model with the best results until now, but unlike Opus, it produced a working result without the need for more than one iteration. Some may argue it was more visually pleasing and had a soundtrack, which was an advantage, but the game’s logic and physics fell short against Opus after a single round of bug fixes. Emerge: The Game, created by Xiaomi MiMo v2 Pro Also, Xiaomi’s model produces these results at a fraction of the cost charged by Anthropic, which could be a major thing to consider for serious projects. The chain-of-thought behavior was different too at first glance. Unlike 4.6, which tucked its reasoning into a separate thinking box (meaning it was not part of the final answer), Opus 4.7 surfaced its chain of thought as part of the main text output. The reasoning was visible and traceable, not hidden behind a UI abstraction, which is a plus for those valuing transparency. Whether Anthropic will keep that behavior or eventually collapse it into a hidden block again is unclear. The token usage was unlike anything we'd seen before. For the first time in our testing, a single session depleted our entire token quota. Watching the model work, we saw it complete a full draft—then write what appeared to be the entire game again from scratch under the label "Rewrite Emerge with bug fixes and improvements," followed by a second pass labeled "Create a rewritten Emerge with bug fixes and improvements." This means, if you’re into serious coding, you’ll be forced to either upgrade your plan, pay a lot on API tokens, or wait a long time until Anthropic resets your usage quotas. Or you could just use a comparable model that charges a lot less Opus 4.6 had never done this. However, it's consistent with what Anthropic warns in the migration guide: more output tokens, especially on agentic tasks at higher effort levels. Opus 4.7 is available today at Claude.ai , the Claude API, Amazon Bedrock, Google Cloud Vertex AI, and Microsoft Foundry. Pricing is unchanged from 4.6: $5 per million input tokens, $25 per million output tokens. Developers can access it via the string claude-opus-4-7. Daily Debrief Newsletter Start every day with the top news stories right now, plus original features, a podcast, videos and more. Your Email Get it! Get it!