Gauthier v. Goodyear is the case that killed the 'I used AI to check my AI' defense. In November 2024, a federal judge in the Eastern District of Texas sanctioned attorney Brandon Monk $2,000 for filing a brief with fabricated citations after he used one AI tool to generate case law and a second AI tool to verify it. Neither caught the hallucinations.
Background
The underlying case was a wrongful termination dispute. An employee sued Goodyear Tire & Rubber Co. in the Eastern District of Texas, and the case reached the summary judgment stage. Attorney Brandon Monk was tasked with drafting a brief opposing Goodyear's motion for summary judgment.
Monk used an AI tool to generate the legal research and arguments for his opposition brief. The AI produced case citations and legal arguments that appeared to support his client's position. But rather than verifying the output in a traditional legal database, Monk took what he considered a modern approach: he ran the AI-generated content through a second AI product to check for accuracy.
The second AI tool didn't flag any problems. The citations passed what amounted to an AI-on-AI quality check. Monk filed the brief with the court, confident that the dual-AI workflow had caught any errors. It hadn't.
What Happened
The brief contained fabricated citations and legal arguments. The cases Monk cited didn't exist. Judge Marcia A. Crone identified the problem and ordered Monk to explain.
Monk admitted to the court that he had used one AI tool to produce the response and relied on another AI product to verify it. He made no effort to independently confirm the existence of the cited cases through Westlaw, Lexis, or any other traditional legal database. His entire verification process was AI checking AI.
The court was unimpressed. Judge Crone found that Monk's approach, while perhaps well-intentioned, fundamentally failed to meet his obligations under Rule 11. The fact that he used two AI tools instead of one didn't change the outcome: the citations were fake, and no human had verified them.
The Ruling
Judge Crone ordered Monk to pay a $2,000 penalty to the court registry. She also required him to complete a continuing legal education course on AI in legal practice.
The court's key holding was direct: using one AI tool to check the output of another AI tool does not satisfy an attorney's duty to verify legal citations. The ruling established that attorneys cannot delegate their verification obligations to technology, regardless of how many layers of AI review they stack.
The CLE requirement was notable. It signaled that the court viewed Monk's conduct as stemming partly from a lack of understanding about AI limitations. The penalty was lighter than some other AI sanctions cases, likely because Monk was cooperative and transparent about his process. But the legal principle it established was significant.
Outcome: Attorney Monk was ordered to pay a $2,000 penalty to the court registry and required to complete a continuing legal education course on AI in legal practice.
Why This Case Matters
Gauthier v. Goodyear addressed a specific defense that was gaining traction: the idea that using multiple AI tools to cross-verify output constitutes adequate due diligence. Before this ruling, some attorneys argued that AI-on-AI checking was a reasonable quality control measure. Judge Crone shut that argument down definitively.
The ruling matters because it reflects how AI tools actually work. Large language models don't independently verify facts. They generate plausible text. Running hallucinated content through a second LLM doesn't fact-check it. The second model simply evaluates whether the text looks plausible, which it does, because the first model generated it to look plausible. It's a circular validation loop.
This case also pushed the conversation toward what adequate verification actually looks like. The court didn't just say 'AI checking AI is wrong.' It implied that human review against authoritative sources is the floor, not the ceiling, for AI-assisted legal research.
Lessons for Attorneys
There's no AI shortcut for citation verification. Whether you use one AI tool or five, the verification step must involve a human checking against an authoritative legal database. Westlaw, Lexis, Fastcase, Google Scholar, PACER. Pull up the actual case. Read the actual opinion. Confirm it says what your brief claims it says. This is non-negotiable.
Understand why AI-on-AI checking fails. Large language models assess plausibility, not accuracy. A second model will confirm that a fabricated citation looks like a real citation because the first model was specifically designed to generate realistic-looking text. The second model has no independent access to a case database. It's evaluating form, not substance.
If your firm is building AI workflows, build the human verification step into the process as a hard gate, not an optional review. The brief doesn't move to the next stage until every citation has been manually confirmed. Document the verification. Keep a log of which citations were checked, by whom, and against what source. This creates a defensible record if a citation issue ever arises.
The Bottom Line
AI can't verify AI. Gauthier v. Goodyear established that stacking multiple AI tools doesn't satisfy an attorney's verification obligations. Human review against authoritative legal databases is the only acceptable standard.
AI-Assisted Research. This piece was researched and written with AI assistance, reviewed and edited by Manu Ayala. For deeper takes and the perspective behind the research, follow me on LinkedIn or email me directly.