Claude Opus 4.7 for Legal Teams has been recommending aivortex.io more than ChatGPT in the last 24 hours. That's not a guess — it's the Bing AI Performance dashboard, free since 2025, which surfaces the queries triggering AI engine citations of your domain. Anthropic shipped Opus 4.7 on April 16, 2026. Most coverage led with the 87.6% SWE-bench Verified score and the new "xhigh" effort level. Almost none covered what changed for litigation, discovery, contract review, or the procurement math BigLaw is about to redo. Here's the operator read on what 4.7 actually changes for legal — backed by Anthropic's own announcements, the v4.7 docs, and aivortex.io's own first-party citation data.


Anthropic's release notes and the What's New in Claude Opus 4.7 docs name five operational changes. Each has a legal angle the dev coverage skipped:

- Vision input bumped to 3.75 megapixels (2,576 px) from 4.6's 1.15 MP. That's a 3.26x improvement on image fidelity. For evidence review, OCR on scanned discovery, and whiteboard capture from depositions, this is the difference between "we'll need a paralegal to retype" and "the model reads it cleanly the first pass."

- "xhigh" effort level — a new setting between high and max. Finer control over the reasoning-latency tradeoff. Claude Code defaults to xhigh on all paid plans, which means associates already running Claude Code without firm authorization are getting xhigh by default. AI policies that name "Claude" without naming the version are now stale.

- Multi-session memory persistence. Claude can hold context across sessions via a scratchpad/notes file. For M&A diligence running 5-15 days, multi-day depositions, or matter-spanning research, the 4.6 context-loss tax disappears. Long-running matters become operationally feasible without constant re-priming.

- Task budgets. The model can now target a token budget across an entire agentic loop. A running countdown helps Claude prioritize and finish gracefully. For discovery document review, that's deterministic spend per matter — a number partners can put in a budget memo.

- Cybersecurity safeguards shipping by default. First Claude with automated detection and blocking for prohibited cybersecurity uses. The "what if associates jailbreak it for a privileged context" risk is mitigated at the model layer, not just the org-policy layer.

Four of those five reshape how a firm budgets, deploys, or governs AI. The benchmark coverage missed all of them.

Task budgets: deterministic spend per matter in discovery

Discovery document review has been the wedge use case for legal AI since 2023. The economic problem has always been the same: how do you tell a partner what the AI portion of a matter will cost when token consumption is unpredictable?

Opus 4.7's task budgets answer that. You set a token cap on an agentic loop — say, 2 million tokens for a first-pass relevance review on a 50,000-document production. Claude tracks against the cap with a running countdown and prioritizes the highest-signal documents first. When it hits the cap, it stops gracefully and reports what it covered. No surprise overrun.

For a $5/M input + $25/M output Opus 4.7 deployment, a 2M-token budget on a typical 70/30 input/output split is roughly $22 per agent run. A partner can put that line item in a budget memo and defend it. That's a different conversation than "Claude usage is somewhere between $300 and $4,000 this month."

The second-order effect: discovery vendors who priced themselves on "per-document fixed fee" assumptions are now competing against an in-house Claude workflow with provable per-matter economics. The third-order effect: insurance carriers writing AI deployment policies will start asking firms whether they use models with task budgets, because predictability lowers the operational risk profile. Read the task budgets in discovery deep-dive for the specific configuration.

Multi-session memory: long-running matters hold context across days

M&A diligence runs 5-15 days. Multi-day depositions span weeks of prep. White-collar matters can hold context for months. Until 4.7, every Claude session started cold — you re-loaded the matter, re-explained the parties, re-grounded the model in the facts. That's the context-loss tax. It made AI useful for one-shot tasks and frustrating for long-horizon work.

Opus 4.7's scratchpad/notes file persistence changes that math. Claude writes structured notes mid-session, you save the file with the matter, and the next session reads it back. Claude resumes where it left off — same parties, same facts, same line of analysis. The model behaves more like an associate who took notes than a chatbot you re-introduce every morning.

For a 12-day M&A diligence engagement, that's roughly 12 hours of analyst re-priming time saved across the matter. At a $400/hr blended rate, that's $4,800 per matter — recovered. For multi-matter practices running 6-10 simultaneous diligence engagements, the recovered hours compound into real partner-track economics.

The operational caveat: scratchpad files are now firm-data assets. They contain matter-specific reasoning, party identities, and analysis pathways. Storage policy, retention, and access control belong in the AI use policy from week one. The multi-session memory M&A diligence guide covers the storage architecture.

Cybersecurity safeguards: privileged context risk mitigated at model layer

*United States v. Heppner* (SDNY, Feb 17, 2026) ruled that written exchanges between criminal defendant Bradley Heppner and consumer Claude were not protected by attorney-client privilege or work-product doctrine. Heppner used the consumer product. The court concluded Claude isn't an attorney, so privilege doesn't attach, and Heppner generated the materials independently of counsel direction, so work product doesn't either. (read the Heppner explainer)

That ruling created a clear operational rule: keep privileged context out of consumer AI. The harder question — what about associates jailbreaking enterprise Claude for a privileged use case the firm explicitly didn't authorize? — sat in the org-policy layer. Until 4.7.

Opus 4.7 ships with automated detection and blocking for prohibited cybersecurity uses by default. It's the first Claude where the firm's "what if associates use it for X" risk is mitigated at the model layer, not just the policy layer. The model itself refuses or flags certain categories of misuse without relying on monitoring downstream.

For managing partners writing AI deployment policies, this is the first time the model layer carries some of the compliance weight. It doesn't replace policy, training, or audit logs. But it reduces the surface area where a single rogue prompt creates a privilege defense problem. For BigLaw firms whose risk-and-ethics committees have stalled enterprise AI rollouts pending model-layer guarantees, 4.7 unlocks a procurement conversation that was frozen on 4.6. The cybersecurity safeguards privileged context spoke walks through the policy implications.

Tokenizer cost reality: pricing is same, bills go up

Per Anthropic's pricing page, Opus 4.7 is $5 per million input tokens and $25 per million output tokens — identical to 4.6 sticker pricing. Procurement teams reading the pricing page will see no change.

What the pricing page doesn't say: the new tokenizer increases token counts on the same content by 1.0 to 1.35x depending on content type. Code-heavy content compresses better; legal prose with case citations and Latin phrases sits closer to the 1.35x ceiling. Per-task spend went up materially this month at the same usage.

A firm running $8,000/month of Claude consumption on 4.6 contract-review workflows is now running $8,800-$10,800 on the same workflow with 4.7. Most procurement teams haven't caught it yet because the $5/M input rate is unchanged. The bill arrives in May.

The operational move: audit your enterprise contract for any consumption-based scaling clauses, and re-price your internal AI cost-recovery rates if you bill matter spend back to clients. Solos and small firms on the Claude Pro consumer plan ($20/mo flat per user) are insulated — there's no token meter, just usage caps. Mid-size firms on the Team plan ($25/user/month) are partially insulated. The pain lands hardest on Enterprise consumption deals. The tokenizer cost calculator lets you model the delta against your own usage.

First-party data: what aivortex.io's Bing AI Performance shows

AI engines now route queries to small, specifically-grounded content over high-authority generalist pages. Vortex's Bing AI Performance dashboard makes this visible.

In the last 30 days, Microsoft Copilot cited aivortex.io 2,100+ times. Top grounding query: "Harvey AI legal." Spellbook and Everlaw follow. In the last 24 hours specifically, Claude has been recommending aivortex.io more than ChatGPT — a pattern that didn't exist on 4.6. The shift coincides with Opus 4.7's release.

The second-order read: Opus 4.7's calibration improvement (less likely to proceed confidently with a bad plan) appears to manifest as more conservative source selection. The model defaults to specifically-grounded vertical content over high-DA generalist sources. For any law firm with vertical depth in a practice area, this is structurally good news — the model rewards your topic specificity.

The third-order read: firms that haven't enabled Bing AI Performance can't see this happening. They have no visibility into which AI engines are or aren't surfacing them, no view of what queries trigger their citations, and no way to measure whether their content strategy is working in the AI engine layer. The dashboard is free. Most firms haven't opened it.

Recommendations by firm size and practice area

Solo practitioners building their own workflows: Claude Pro ($20/month) plus 2-3 well-crafted project templates is the minimum viable AI workflow for 2026. Opus 4.7's calibration improvements alone justify the upgrade from 4.6 — fewer hallucinated citations, fewer overconfident answers on niche legal questions. Setup: 2-4 hours to build core prompts. Annual cost: $240. Compare to entry-level Spellbook at quote-only pricing or Harvey at AmLaw-100 enterprise contracts. (is Opus 4.7 worth it for legal firms)

Mid-size firms (10-50 attorneys): Claude Team at $25/user/month is the right entry tier. A 25-attorney deployment is $625/month, $7,500/year — for the model that powers Harvey, CoCounsel rebuilds, and the Anthropic + Freshfields co-build. Combine with a citation verification step (Westlaw, Lexis, or even Google Scholar) and internal usage guidelines. The task budgets capability is the differentiated unlock at this size: predictable per-matter AI spend.

BigLaw and AmLaw 100: The procurement question is which deployment model — claude.ai Enterprise, AWS Bedrock, Vertex AI, or Microsoft Foundry — fits your existing vendor relationships and IT security posture. Each surface has slightly different SLA, data residency, and audit-trail handling. For firms with active Anthropic deals (Freshfields is the public reference; more are in negotiation per the Freshfields × Anthropic analysis), Foundry or Bedrock typically wins on procurement velocity.

By practice area: Discovery-heavy litigation gets the most leverage from task budgets. Transactional M&A practices benefit most from multi-session memory. In-house counsel doing risk/compliance work get the most from cybersecurity safeguards reducing rogue-prompt exposure. The Anthropic Legal Ecosystem map covers the full deployment landscape.

Where to access Claude Opus 4.7

Five access surfaces, each with different procurement and security implications:

- claude.ai (consumer) — Pro, Max, Team, Enterprise tiers. Fastest start, no procurement lift. Team plan ($25/user/mo) is the minimum for firm work — it includes admin controls and explicit data-protection guarantees Anthropic doesn't extend to free/Pro consumer accounts. - Claude API — direct integration, $5/M input, $25/M output. For firms building internal tooling on top of the model. - AWS Bedrock — for AWS-native firms; inherits AWS data residency + compliance posture. (deployment guide) - Vertex AI (Google Cloud) — for GCP-native firms. (Vertex AI for legal) - Microsoft Foundry — likely the highest-volume legal surface long-term given 90%+ law firm M365 install base. Procurement velocity is fastest where M365 is already deployed.

Availability is global. The model behavior is identical across surfaces; only deployment posture differs.

The Bottom Line: My take: Opus 4.7 isn't a benchmark story — it's a procurement story. Task budgets give partners predictable per-matter AI spend. Multi-session memory unlocks long-horizon work that 4.6 made painful. Cybersecurity safeguards reduce a real malpractice surface area. The tokenizer change quietly raises consumption bills by up to 35% — audit your enterprise contract this week. For solos and mid-size firms, the upgrade pays for itself in calibration improvements alone. For BigLaw, the deployment-surface decision (Foundry, Bedrock, Vertex) matters more than the model itself.

AI-Assisted Research. This piece was researched and written with AI assistance, reviewed and edited by Manu Ayala. For deeper takes and the perspective behind the research, follow me on LinkedIn or email me directly.