Lexis+ AI is the most accurate AI legal research tool available in 2026, scoring 65% on Stanford's 2025 Legal AI Benchmark -- but that means it's wrong more than a third of the time. Westlaw Precision with CoCounsel scored 42% on the same benchmark, and no tool tested above 70% on complex multi-jurisdictional queries.
These numbers aren't meant to scare you away from AI research -- they're meant to calibrate your expectations. AI legal research is dramatically faster than manual methods, but 'faster' and 'accurate' aren't the same thing. Every tool requires a verification workflow, and the firms getting real value from AI research are the ones that built that workflow into their process from day one.
The Stanford 2025 Benchmark: What Was Actually Tested
Stanford's Legal AI Benchmark (published late 2025) tested AI research tools across five categories: case retrieval accuracy, statutory interpretation, regulatory analysis, multi-jurisdictional research, and brief drafting assistance.
The results, ranked by overall accuracy: 1. Lexis+ AI: 65% overall accuracy 2. vLex Vincent AI: 58% (strongest on international/comparative law) 3. Harvey: 53% (strongest on firm-knowledge integration) 4. CoCounsel (Casetext/Thomson Reuters): 49% 5. Westlaw Precision: 42% overall accuracy
Thomson Reuters publicly disputed the methodology, arguing that the benchmark overweighted complex edge cases. They're partially right -- on straightforward single-jurisdiction queries, both Westlaw and Lexis tools perform significantly better (75-85% accuracy). The gap emerges on complex research tasks: multi-jurisdictional analysis, novel legal questions, and regulatory interpretation.
The hallucination rates tell an equally important story. Lexis+ AI hallucinated citations in under 3% of responses (down from ~12% in early 2024). Westlaw Precision hallucinated at 5-6%. Harvey claims sub-2% but hasn't submitted to independent testing. Every tool still makes things up -- the question is how often, and whether your workflow catches it.
Why Accuracy Numbers Don't Tell the Full Story
A 65% accuracy score doesn't mean Lexis+ AI gives you the wrong answer 35% of the time. Accuracy in the Stanford benchmark measures multiple dimensions:
Citation accuracy (did the case exist and say what the tool claimed): All tools score above 94% here. Hallucination is a real problem but it's gotten dramatically better.
Relevance ranking (did the tool surface the most on-point authority first): This is where big differences emerge. Lexis+ AI puts the best case in the top 3 results 78% of the time. Westlaw Precision does it 61% of the time. For a researcher, this is the difference between finding your answer in 5 minutes and finding it in 30.
Completeness (did the tool miss important cases): The toughest category. All tools miss relevant authority -- the question is whether they miss *the* case that would change your analysis. On this measure, no tool scored above 60%.
The practical takeaway: AI tools are excellent at getting you to the right neighborhood fast. They're unreliable at guaranteeing you haven't missed something. Use AI for speed, use traditional verification for completeness.
Head-to-Head: Lexis+ AI vs. Westlaw Precision
The two platforms most firms choose between:
Lexis+ AI strengths: - Higher accuracy on complex queries (65% vs. 42%) - Better natural language understanding -- you can ask conversational questions - Stronger case summarization and analysis - Lower hallucination rate (under 3%) - Better integration with Lexis+ research database
Westlaw Precision strengths: - Superior statutory and regulatory coverage - Better citation verification tools (KeyCite still leads Shepard's for some practitioners) - Stronger transactional law databases (Practical Law integration) - More robust Boolean search fallback when AI fails - Better for experienced researchers who know exactly what they're looking for
Pricing: - Lexis+ AI: $150-250/user/month depending on firm size - Westlaw Precision with CoCounsel: $200-350/user/month
The verdict: Lexis+ AI is more accurate for AI-first research workflows. Westlaw Precision is stronger as a hybrid platform where you use both AI and traditional search. If your firm's associates primarily use natural language queries, Lexis wins. If they're experienced researchers who use AI to supplement Boolean searching, Westlaw may be the better fit.
The Verification Workflow Every Firm Needs
No matter which tool you choose, you need a verification workflow. Here's what the best firms are doing:
Step 1: AI query. Run your research question through the AI tool. Get the initial case list, analysis, and recommended authority.
Step 2: Citation verification. Check every case cited by the AI. Confirm it exists, confirm the citation is correct, confirm the holding matches what the AI claimed. This takes 2-3 minutes per case -- non-negotiable.
Step 3: Shepardize/KeyCite. Verify every case is still good law. AI tools are inconsistent at flagging overruled or distinguished authority. Don't skip this.
Step 4: Completeness check. Run a traditional Boolean search on the key legal issue to catch cases the AI missed. This is where AI saves you time -- you're not doing the full search, just a targeted check.
Step 5: Document the workflow. Note in your research memo that AI was used for initial research and all citations were independently verified. This protects you if questions arise.
Total time: AI query (5 min) + verification (20-30 min) vs. traditional research (60-90 min). The time savings are real -- 40-60% -- even with thorough verification. But only if you actually do the verification.
Beyond the Big Two: Alternative Tools Worth Considering
vLex Vincent AI scored 58% on Stanford's benchmark and offers the best international and comparative law research capability. If your practice involves multi-jurisdictional work, vLex at $79/user/month is a strong complement to either Lexis or Westlaw.
Harvey isn't a traditional research tool -- it's a reasoning layer that integrates with your firm's knowledge base. It scored 53% on general legal research but excels at finding relevant precedent within your own firm's work product. At $150/user/month with minimum seat requirements, it's an Am Law play.
Bloomberg Law's AI assistant is strong for transactional, regulatory, and financial law research where Bloomberg's data advantage matters. It didn't participate in the Stanford benchmark but practitioners report accuracy comparable to Westlaw Precision.
Google Scholar + AI workflow: Some practitioners use general-purpose AI (Claude, GPT) for initial analysis, then verify citations in Google Scholar (free) and Lexis/Westlaw (paid). This is the budget approach -- less reliable but costs nothing beyond the AI subscription.
The Bottom Line: Lexis+ AI leads accuracy at 65% on Stanford's 2025 benchmark, with Westlaw Precision at 42%. No tool is reliable enough to use without citation verification -- every firm needs a verification workflow that adds 20-30 minutes but catches the 3-6% hallucination rate. The real value of AI research isn't accuracy alone, it's the 40-60% time savings when you combine AI speed with human verification.
AI-Assisted Research. This piece was researched and written with AI assistance, reviewed and edited by Manu Ayala. For deeper takes and the perspective behind the research, follow me on LinkedIn or email me directly.
