OpenAI released GPT Image 2 on April 21, 2026. 4K resolution. Roughly 99% character-level text accuracy across Latin, CJK, Hindi, and Bengali scripts. Up to 16 reference images in a single prompt. A reasoning pipeline that self-checks outputs before returning them. The technical achievement isn't what matters for litigation. The provenance question is. A federal judge will see an AI-generated demonstrative aid in evidence within 90 days, and the Federal Rules of Evidence don't yet have a provenance standard. Vortex tracks 300+ federal judges with AI standing orders and the 1,227 documented hallucination sanctions in Damien Charlotin's database. Image-generation will be the next category these orders try to catch. Most of them will catch it badly. Per OpenAI's Images 2.0 announcement, the model is now in the API and Codex by default — meaning any associate with API access can produce courtroom-grade visuals before lunch. Here's the operator read on what changes for evidence, and what firms need to do this week.
What actually shipped on April 21, 2026 — and why it's an evidence problem, not a design problem
Per OpenAI's Images 2.0 product page and the gpt-image-2 developer announcement, GPT Image 2 ships five capabilities that move it from "AI art toy" to "forensic-grade output device":
- 4K resolution — courtroom display-grade. Demonstratives no longer look like AI artifacts at 30 feet on a projector. - ~99% character-level text accuracy across Latin, CJK, Hindi, Bengali scripts. The 2024-2025 "AI can't render text" tell is gone. Fake exhibits, fake signage, fake email screenshots, fake document renders all become plausible. - Up to 16 reference images as input. An attorney can feed 16 photos of a real intersection and get a synthetic image that matches lighting, signage, and angle. - Multi-turn editing preserving context. Iterative refinement of the same scene, same characters, same room — across an entire matter prep cycle. - Reasoning pipeline self-checks outputs. The model flags inconsistencies in its own renders before delivering them. Detection signals that adversarial parties used to rely on are now scrubbed by the model itself.
The second-order read: the technical bar for "is this image real?" just collapsed. The third-order read: every detection technique that depended on rendering artifacts (six fingers, broken text, lighting impossibilities) is now obsolete. Forensic image authentication will need to move from artifact-detection to provenance-attestation — and that's a different stack entirely.
The deeper problem isn't the model. It's the distribution. GPT Image 2 sits inside ChatGPT Plus, Pro, Business, Enterprise, the API, and Codex. Roughly 90%+ of US litigators have a ChatGPT account. Roughly zero firms have a written policy for what associates can and can't generate.
The Federal Rules of Evidence weren't built for this
Federal Rule of Evidence 901 governs authentication: a proponent must produce evidence "sufficient to support a finding that the item is what the proponent claims it is." Rule 902 lists 14 categories of self-authenticating evidence — items that don't require extrinsic proof. In 2017, Rule 902 was amended to add subsections (13) and (14) covering electronic records authenticated by a qualified person's certification. That's the most recent textual update directly applicable to digital evidence per the LII text of Rule 902.
Neither subsection contemplates synthetic images. Both presume the underlying record is a *true copy* of something that existed in the physical or digital world. AI-generated images are not copies. They're new objects with no physical-world referent.
The Advisory Committee on Evidence Rules has had AI-authentication on its agenda since 2024. The committee's April 2024 report noted the issue and deferred. Two years later, the rules still haven't moved. Meanwhile, 300+ federal judges have written their own AI standing orders: most addressing text generation, almost none addressing image generation. The Ropes & Gray AI Court Order Tracker confirms the gap.
The operational read: until the rules catch up, authentication objections to AI-generated visuals will be litigated case-by-case under 901's general standard. That's a procedural mess for the next 18-24 months. The federal rules of evidence 902 deep-dive walks through how the current text actually applies. And where the gaps create exposure.
The second-order read: a smart trial team will not wait for the rules to catch up. They'll build a private chain-of-custody protocol that survives a 901 challenge regardless of what the rules eventually say. The third-order read: the firms that build that protocol first turn it into a procurement asset; "we use AI for prep, with documented provenance" sells better to risk-conscious clients than "we don't use AI."
Provenance: C2PA Content Credentials is the standard the courts will land on
C2PA, the Coalition for Content Provenance and Authenticity: published its first technical specification in 2022 and reached v2.x by late 2025. Per the C2PA technical specifications, the standard embeds cryptographically signed metadata in image files describing how the asset was created, who created it, and what edits were applied. The consumer-facing implementation is Content Credentials (contentcredentials.org), backed by Adobe, Microsoft, OpenAI, Leica, Sony, and others.
OpenAI added C2PA Content Credentials to DALL-E 3 outputs in February 2024 per Adobe's Content Credentials integration timeline. GPT Image 2 inherits and extends this. Every image generated through the API or ChatGPT carries embedded C2PA metadata identifying it as AI-generated.
The operational implication for evidence: a smart litigator can now inspect any image's C2PA manifest before stipulating to authentication. The bad news: the metadata is strippable. A re-export, a screenshot, a re-save through any editor that doesn't preserve the manifest, or a deliberate scrub all break the chain. The standard is voluntary. There's no legal requirement to preserve it.
The second-order read: federal courts will eventually require C2PA preservation in discovery production protocols. The 2024 amendments to FRCP 26 are the closest analog; "reasonably accessible" electronically stored information now arguably extends to embedded metadata. The third-order read: discovery production vendors who today strip metadata to reduce file size will face FRCP-grade exposure once the first sanctions opinion calls it spoliation. Read the C2PA Content Credentials evidence standards spoke for the full preservation playbook.
My take: every firm should adopt a discovery production protocol this quarter that explicitly preserves C2PA metadata on any image asset produced or received. That's a one-paragraph policy update with a years-of-litigation moat.
Demonstratives, depositions, discovery, three places this lands first
GPT Image 2 enters litigation through three doors before the rules update. Each door has a different exposure profile.
Door 1: Demonstrative aids. A trial team prepares a demonstrative: a recreated intersection, a reconstructed crime scene, a simulated email thread, a rendered product defect. Today, the demonstrative is built by an outside graphics vendor and disclosed under the local pretrial order. With GPT Image 2, an associate builds it in 20 minutes from 16 reference photos. The disclosure question gets harder. Most federal local rules require disclosure of demonstratives in advance of trial. None specify whether AI-generated demonstratives need separate disclosure as AI-generated. The AI demonstratives courtroom disclosure rule gap analysis covers the jurisdictional split.
Door 2: Deposition exhibits. A deposing attorney shows a witness an image. "is this a fair and accurate representation of the location on the date in question?" If the image is AI-generated and the witness affirms, the resulting testimony enters the record under a corrupted authentication. The deposition transcript doesn't carry C2PA metadata. The exhibit attached does; if anyone bothers to check. The deposition exhibits AI image disclosure analysis walks through the witness-prep angle.
Door 3: Discovery production. A party produces documents containing images, photos, screenshots, scans. The producing party doesn't know which images in their custodian's pool were AI-generated. The receiving party doesn't either. With GPT Image 2 generating photo-realistic outputs at 4K, neither side has a quick way to triage. Detection tools lag generation tools by 6-18 months structurally. The how to detect AI-generated images in discovery production guide covers the current toolkit.
Four additional vectors will surface within 12 months: damages exhibits (synthetic injury photos), expert witness illustrations (synthetic methodology diagrams), settlement-leverage exhibits (synthetic scene reconstructions), and FOIA / public records production (synthetic government records). Firms should start building authentication and detection muscle now: the demand will outrun supply hard.
The first sanctioned attorney is coming. Here's the structural prediction
The pattern is predictable from the sanctions data. Per the Charlotin AI Hallucination Database, text-generation sanctions went from zero in 2022 to 1,227 documented globally by early 2026. Acceleration: roughly 5-6 new documented cases per day. Court reactions tracked the same shape across jurisdictions: novelty period (~6 months), early sanctions (~12 months), pattern recognition (~18 months), local rules (~24 months).
Image generation is on month 1 of that cycle. Apply the same structural factors and the prediction is direct: a federal court will sanction an attorney for a demonstrative aid or deposition exhibit produced by GPT Image 2 (or competitor) within 90 days of the model's broad availability. The factors:
1. Availability is now broad. API + Codex + ChatGPT Plus puts GPT Image 2 in front of millions of attorneys. 2. Detection is asymmetric. Generation is faster than detection by structural design. Adversarial parties can produce faster than opposing counsel can verify. 3. Disclosure rules are vacant. No mandate to disclose AI-generation in most local rules means no clean violation, but a 901 authentication failure will fill the gap. 4. The first case is high-publicity. Every sanctioned-attorney-uses-AI case in 2023-2026 made national news. Image-generation cases will too; judges know this and will write strong opinions. 5. Insurance carriers are watching. Sanctions exposure prices into malpractice premiums within 90 days of pattern emergence.
The first sanctioned attorney AI image prediction analysis walks through the case-fact pattern most likely to surface first. This is structural analysis, not naming names, but the trial team that doesn't have a written AI-image policy by Q3 2026 is taking on uncompensated risk.
What firms should do this week: the policy moves before the rules catch up
Five operational moves that don't require waiting for FRE updates or state bar opinions. All are inside-the-firm policy work.
Move 1: Adopt a written AI-image policy. Two pages. Define what's allowed (demonstratives flagged as AI-generated, with C2PA preserved), what's prohibited (any AI image presented as a photographic record without disclosure), and the disciplinary path for violations. The firm policy template for AI-generated images in evidence prep provides a structural framework. Don't wait for ABA model rules.
Move 2: Update litigation hold and discovery protocols. Add AI-generated content to the standard hold language. Add C2PA metadata preservation to the production protocol. Add an explicit "AI-generation detection" sub-step to inbound document review.
Move 3: Train deposition-taking attorneys. New questions for witness prep: "How was this image obtained? Was any AI tool used in its preparation? Does it carry Content Credentials?" Build it into the standard witness prep checklist.
Move 4: Inventory which AI tools associates currently use. Most firms can't answer this. Run a 30-day usage audit. Then categorize: prohibited, conditional, approved with disclosure. Update the engagement letter language so clients understand the firm's AI posture.
Move 5: Subscribe to the state bar opinion stream. State bars are starting to issue ethics opinions on AI image generation specifically. California Bar Formal Opinion 2025-something is rumored to be in development. Track the state bar ethics opinions on AI image generation analysis for the running list.
The second-order read: the firms that ship these five moves first will be the firms that win the first "our adversary used AI without disclosure" motion when the inevitable case lands. The third-order read: this is one of the rare 2026 windows where policy work compounds faster than litigation work. The motion practice is downstream of the policy infrastructure. Build the infrastructure first.
The vendor question; GPT Image 2 vs Midjourney vs Flux for legal use
Three primary models are in production litigation prep workflows today: GPT Image 2 (OpenAI), Midjourney v7 (Midjourney Inc.), and Flux Pro 1.1 (Black Forest Labs). Each has legitimate use cases. None is the wrong answer in the abstract. The fit decision depends on what the firm is producing and how rigorously it documents provenance.
GPT Image 2 ships C2PA metadata by default per OpenAI's Images 2.0 announcement, integrates with the broader ChatGPT enterprise stack, and offers reasoning-pipeline self-checks. It fits firms already on ChatGPT Business or Enterprise, same procurement, same data handling, same audit trail. The tokenizer means an image generation costs comparable to ~$0.05-0.20 per render at API rates per the OpenAI API pricing page.
Midjourney v7 ships through Discord and a web portal. It produces best-in-class aesthetic results for non-photorealistic concepts. It does not ship C2PA metadata by default. For courtroom demonstratives, the lack of provenance metadata is a meaningful procurement objection.
Flux Pro 1.1 is Black Forest Labs' model, available via fal.ai, Replicate, and direct API. Open weights make it deployable in private cloud environments: relevant for firms with strict data-residency posture. C2PA support is partial through implementer choice.
The fit summary, not character: For courtroom-bound work, GPT Image 2's C2PA-by-default is the right starting point. For brand and marketing assets, Midjourney's aesthetic edge wins. For high-confidentiality on-prem deployment, Flux's open weights win. The GPT Image 2 vs Midjourney vs Flux legal disclosure comparison covers the per-use-case math in detail. None of these vendors is the villain. The villain is the firm that doesn't have a policy.
First-party data: what Vortex's coverage of court orders says about what's coming
Vortex maintains coverage of every federal district's AI standing order at aivortex.io/legal/ai-disclosure/. 94 district pages, updated against the Bloomberg Law Federal Court AI Standing Orders comparison and the Ropes & Gray tracker. The pattern in the 2023-2025 wave of orders is direct: when a single judge in a district issues an AI order, neighboring judges issue similar orders within 60-90 days. By month 6, a majority of the district has them. By month 12, the Bloomberg Law tracker has the district fully mapped.
What the tracker shows for image generation: almost no current standing order names AI-generated images specifically. The orders address "generative AI" or "AI-assisted drafting"; language that arguably covers images but doesn't say so. The first judge to write an image-specific order will trigger the same 60-90-day cascade. Vortex's first-party data on which districts cite Vortex pages most heavily (top three: SDNY, NDIL, NDTX per Bing AI Performance) gives an early read on which districts are most actively researching AI policy and likely to write the first image-specific order.
The second-order read: firms with active matters in these districts should brief the bench on AI-image authentication standards proactively, before the standing order writes itself in a way that disadvantages the firm's clients. The third-order read: the standing-order text of the first image-specific order will become the de-facto national template, same way Judge Brantley Starr's 2023 NDTX order became the template that 50+ subsequent judges copied with minimal modification.
Where to access GPT Image 2: and which surface fits which firm
Five access surfaces, each with different procurement and provenance implications:
- ChatGPT Plus / Pro / Business / Enterprise. Fastest start, no procurement lift on top of existing ChatGPT seats. Per the ChatGPT pricing page, Plus is $20/month, Pro is $100 or $200/month depending on tier, Business is $20-25/user/month, Enterprise is quote-only. Business and Enterprise carry data-handling commitments that Pro and Plus don't. For privileged matter work, Business is the minimum; Plus and Pro are consumer surfaces. - OpenAI API direct, $0.05-0.20/image roughly, depending on resolution and reasoning-pipeline depth. Best for firms building internal tooling on top of generation. Carries full enterprise data-handling commitments under the API DPA. - Codex (the OpenAI developer environment): ships gpt-image-2 by default per the developer announcement. Fits legal-tech engineering teams building internal demonstrative tools. - Microsoft Foundry. The Azure-hosted OpenAI model surface. Fits firms with M365 deployment posture and existing Azure procurement. Same data-handling guarantees as Microsoft Copilot for M365. - Third-party wrappers (Poe, OpenRouter, etc.); variable data-handling. Generally not appropriate for privileged-matter use unless the wrapper carries an explicit DPA.
The procurement decision tree: privileged work → Business / Enterprise / API / Foundry. Marketing and non-privileged work → any surface, with C2PA preservation as the only hard requirement. The Microsoft Foundry legal procurement guide covers the M365 deployment posture in detail.
The Bottom Line: My take: GPT Image 2 isn't a design story for legal, it's an evidence story. The Federal Rules of Evidence don't yet have a provenance standard, and a federal sanction in this category will land within 90 days of broad availability. The firms that ship a written AI-image policy this quarter, adopt C2PA metadata preservation in discovery production, and update deposition prep to ask provenance questions are buying years of motion-practice insulation for two pages of policy work. The vendor decision is secondary. The policy decision is primary.
AI-Assisted Research. This piece was researched and written with AI assistance, reviewed and edited by Manu Ayala. For deeper takes and the perspective behind the research, follow me on LinkedIn or email me directly.
