Anthropic ran an experiment four days ago. 69 employees. $100 budgets. 186 completed deals. The agents handled the buying, the selling, and the negotiation. Lawyers weren't in the room. Per Artificial Lawyer's April 27, 2026 coverage, 500 items were listed across the marketplace and total transaction value cleared just over $4,000 in a single internal pilot. Legal IT Insider led with the part that matters for firms: "the legal frameworks don't exist." That's the gap firms haven't started solving, and the part the LinkedIn coverage skipped. Here's the operator read on Project Deal, what it actually demoed, and the supervision-and-liability stack legal needs to build before the second pilot ships.


What Project Deal actually was: the numbers, not the narrative

Anthropic recruited 69 employees on April 24, 2026, gave each $100 in buying budget, and let Claude agents represent the buyers and sellers in an internal marketplace. The agents listed items, negotiated terms, completed transactions, and handled dispute escalation between themselves. Outcome per Artificial Lawyer: 186 completed deals across roughly 500 listings, total transaction value just over $4,000.

The naive read is "toy numbers, ignore it." That's the wrong read. The pilot wasn't designed to move money. It was designed to test whether agents can negotiate, transact, and resolve disputes without humans approving each step. They can. The ratio (186 deals from 69 humans giving budget instructions, mediated by agents) is the part that matters. Agents are the transactional layer. Humans set the budget envelope.

The second-order read: every consumer-marketplace pattern that took 15 years to build (eBay's feedback system, Amazon's A-to-Z guarantee, Stripe's chargeback flow) is about to be re-litigated for agent-to-agent flows in compressed time. The third-order read: the firm that writes the first defensible "agent transaction supervision" engagement letter wins the next 24 months of inbound from corporates running these pilots internally. None of the AmLaw 100 has shipped that template yet.

Legal IT Insider's headline is the right frame: "the legal frameworks don't exist." Specifically, four gaps:

- Agency law. The Restatement (Third) of Agency § 1.01 defines an agent as a "person" acting on behalf of a principal. Comments contemplate corporate agents, not software agents. Whether an LLM agent qualifies as an "agent" with the duties Restatement § 8.01-8.11 imposes (loyalty, care, disclosure) is unsettled. No appellate court has reached the question on facts like Project Deal's. - UCC Article 2. Sale-of-goods formation rules (offer, acceptance, consideration) presume human or corporate parties. UCC § 2-204 lets contracts form "in any manner sufficient to show agreement," which arguably covers agent-mediated formation. Enforceability against the principal when the agent goes off-script is open. - Model Rules of Professional Conduct. ABA Model Rule 5.3 covers supervision of nonlawyer assistants. The 2024 ABA Formal Opinion 512 on generative AI extended Rule 5.3 to AI tools used by lawyers. Project Deal scenarios where the AGENT itself is the principal's representative, not the lawyer's tool, sit outside that opinion. - Insurance and bonding. No standard malpractice carrier currently rates agent-supervision risk separately from general AI use. That changes within 12 months. See the escrow and arbitration gap analysis for the bonding angle.

The operator move: don't wait for ABA or NCCUSL to draft the framework. The firms billing on this in 2027 are writing the engagement-letter language now.

Supervision: which Model Rule applies when an agent acts for a party

ABA Formal Opinion 512 (July 2024) extended Model Rule 5.3 supervision duties to AI used as a tool by the lawyer. Project Deal flips the relationship. The AI agent is acting as the party, not as the lawyer's assistant. That's a different supervision question.

The closest analog is Model Rule 1.2(c) (limitation of representation scope) combined with Rule 1.4 (communication with client). If a corporate client deploys agents to transact and engages a firm to oversee the agent layer, the engagement scope has to specify: which transactions the lawyer reviews, the agent's authority envelope, escalation triggers, and audit-log retention. None of those are standard engagement-letter clauses today.

The second-order issue is Rule 1.1 competence. A managing partner who agrees to supervise agent transactions without understanding token budgets, prompt injection risk, or the model's calibration profile is in the same posture as a partner who agreed to supervise complex e-discovery in 2010 without understanding TAR. Competence-by-association ages badly. ABA Formal Opinion 512 already requires "reasonable understanding" of generative AI capabilities. A future opinion on agent supervision is foreseeable.

The third-order issue is Rule 5.5 unauthorized practice of law. If the agent is drafting transaction documents, negotiating terms, and binding the principal across state lines, the geographic-licensing analysis is non-trivial. The agent has no jurisdiction. The principal does. The lawyer supervising has yet another. See the agent supervision rules deep-dive for the per-state breakdown the framework needs.

Fiduciary duty: when an agent represents both sides

Project Deal's setup let one party's agent negotiate against another party's agent. In some completed deals, the same underlying model architecture (Claude) sat on both sides. That's a structural problem traditional agency law has dealt with for centuries. Dual representation requires informed consent of both principals.

For human agents, the consent flow is documented (Restatement § 8.06, conflict-of-interest disclosure). For software agents, no equivalent exists. The principal granted the agent budget authority. Did the principal consent to the agent transacting against another instance of the same model? Did the principal even know? Most users don't read the model card.

The second-order problem: information asymmetry inside the model. Two Claude instances negotiating against each other share training data, architectural priors, and reasoning patterns. They're not independent counterparties in any meaningful sense. The textbook cartel concern (parallel pricing without explicit collusion) gets a new shape when the "parallel" comes from shared model weights, not shared phone calls.

The third-order problem: enforcement gap. The FTC has guidance on algorithmic collusion (per Khan-era statements). Whether two agent instances behaving cooperatively because they share a model counts as "agreement" under Sherman Act § 1 is genuinely open. The framework needs a rule, not a vibe. Read the AI agent fiduciary duty analysis for the dual-representation doctrine map.

Heppner meets Project Deal: privilege when agents talk

*United States v. Heppner* (SDNY, Feb 17, 2026, Judge Rakoff) ruled that written exchanges between criminal defendant Bradley Heppner and consumer Claude were not protected by attorney-client privilege OR work-product doctrine. Claude isn't an attorney, so privilege doesn't attach. Heppner generated the materials independently of counsel direction, so work product doesn't either. (read the full Heppner explainer)

Project Deal generalizes that problem. Every agent-to-agent transaction generates a written record: prompt, response, negotiation log, decision rationale. Under Heppner's reasoning, none of those communications carry privilege. They're discoverable.

For a corporate principal running internal agent pilots, that's a documentation explosion in any future litigation. Discovery requests will ask for the full agent transcript. The transcript will include the principal's strategic preferences, walk-away points, and counterparty research, all unprivileged.

The second-order effect: in-house legal teams that were comfortable with consumer Claude for low-stakes brainstorming need a different posture for agent-deployment. The Team or Enterprise tier with explicit data-handling commitments matters more than ever. The third-order effect: insurance carriers writing cyber liability for agent deployments will start asking for the deployment surface (consumer vs Team vs Enterprise vs Bedrock vs Foundry) at the underwriting stage. Most firms haven't put that in their renewal packet. The Heppner-meets-Project-Deal privilege spoke walks through the documentation architecture.

Escrow, arbitration, dispute resolution: who arbitrates between agents?

Project Deal's 186 deals didn't all close cleanly. Some required dispute resolution between the buyer's agent and seller's agent. The resolution was handled inside the agent layer. The model adjudicated.

For a $4,000 internal pilot, that's fine. For a $4M B2B agent transaction in 2027, it isn't. Federal courts, AAA arbitration, JAMS: none have rules tailored to disputes where the parties are agents acting under principal-set budget envelopes. The Federal Arbitration Act presumes human signatories.

The practical short-term answer is escrow with human checkpoint. A dispute over $X triggers a human review by counsel for the principal, with the agent's transcript as the evidentiary record. The escrow provider holds funds until the checkpoint resolves. That's a workflow firms can build now, not a regulatory waiting game.

The second-order question is enforceability across jurisdictions. If a Delaware-incorporated principal's agent transacts with a Texas counterparty's agent and the dispute is resolved in the agent layer, where is the resolution enforceable? The answer is currently "nowhere reliably." The first firms to draft choice-of-law-and-forum clauses tailored to agent transactions own the procurement conversation when corporates start asking. See the escrow and arbitration framework for the clause stack.

Billable hours: should firms charge for AI-supervised transactions?

The naive answer is no. The AI did the work. The sophisticated answer is firms should charge for the supervision architecture they design, not the transaction-by-transaction labor.

For Project Deal-style flows, the firm's value is up-front: the engagement letter, the authority envelope, the escalation triggers, the audit-log retention policy, the dispute-resolution clause. That's measurable as a fixed-fee deliverable per agent deployment, analogous to a SaaS contract review or a vendor-procurement playbook.

The second-order economics shift the partner-track math. A 25-person mid-market firm that designs three agent-supervision frameworks per quarter at a fixed fee per framework generates billable revenue without the per-matter hourly grind. That's a different practice shape than commodity transactional work.

The third-order effect is on insurance. Carriers will start pricing agent-supervision malpractice as a separate line item by the 2027 renewal cycle. Firms that document their supervision architecture (via engagement letters and audit logs) will pay less. Firms that wing it will pay more or be excluded. The billable hours and AI-supervised work analysis covers the fee-structure math.

Recommendations by firm size and practice area

Solo and small-firm transactional practices: Don't compete on volume against agent-mediated marketplaces. Compete on the supervision layer. Offer a fixed-fee "agent deployment review" for small businesses running internal pilots: engagement letter, authority envelope, audit-log policy, escrow language. Annual revenue potential per client: $2,500-7,500. Setup time: build one template, reuse across clients.

Mid-market firms (10-50 attorneys): The procurement opportunity is corporate clients running 5-50 internal agent pilots and needing a firm to write the supervision framework once and audit it quarterly. That's a fixed-fee retainer model with predictable margin. The firms that ship a defensible template in the next 90 days own the procurement conversation when these pilots scale.

BigLaw and AmLaw 100: The opportunity is the playbook, not the per-matter work. Anthropic, OpenAI, Google, and the agent-deployment vendors will need a small number of firms to build the canonical agent-transaction-governance framework. The first firm to publish a defensible whitepaper plus engagement-letter template plus state-by-state supervision analysis owns the next decade of inbound. Freshfields' Anthropic co-build (covered in the Anthropic legal ecosystem map) is the precedent: co-build with a foundation model rather than wait for the spec.

By practice area: M&A practices benefit from agent-supervised diligence (faster, cheaper, traceable). Commercial litigation practices benefit from the discovery explosion (every agent transcript is new evidence). Regulatory practices benefit from the framework-drafting opportunity. Insurance defense practices benefit from the new line of agent-supervision malpractice claims that will emerge by 2027.

The Bottom Line: My take: Project Deal isn't a tech demo. It's the canary for transactional law without lawyers in the room. The 69-employee, $100-budget, 186-deal pilot proved agents can transact end-to-end. It also proved the legal framework underneath has gaps measured in years, not months. Firms that ship a defensible agent-supervision template in the next 90 days will be in the room when corporates scale these pilots. Firms that wait for the ABA to opine will be billing commodity work in 2028.

AI-Assisted Research. This piece was researched and written with AI assistance, reviewed and edited by Manu Ayala. For deeper takes and the perspective behind the research, follow me on LinkedIn or email me directly.