TwinLadder logoTwinLadder
TwinLadder
TwinLadder logoTwinLadder
Back to Archive
TwinLadder Intelligence
Issue #16

TwinLadder Weekly

September 2025

TwinLadder Weekly

Issue #16 | September 2025


AI Hallucination Cases Database: Global Sanctions and Lessons

Lawyers worldwide are being sanctioned for AI hallucinations. Don't be next. Here's what Damien Charlotin's database reveals about patterns, sanctions, and prevention.


The Growing Record

Damien Charlotin, an independent practitioner at Pelekan Data Consulting and Research Fellow at HEC's Smart Law Hub, maintains what has become the definitive tracker of AI hallucination cases in legal proceedings worldwide.

The numbers are sobering.

As of late October 2025, the database catalogues 486 cases globally, with 324 in the United States alone. The growth rate has accelerated: "Before this spring in 2025, we maybe had two cases per week," Charlotin told Cronkite News. "Now we're at two cases per day or three cases per day."

This isn't a US phenomenon. The database includes cases from Canada, the UK, Australia, and multiple other jurisdictions. The common thread: lawyers trusting AI-generated content without verification.

The Anatomy of AI Hallucinations

Charlotin's research identifies three distinct failure modes:

Type Description Detection Difficulty
Fabricated cases AI invents non-existent case citations Easy (case doesn't exist)
Fake quotes Real case, fabricated quotation Medium (case exists, quote doesn't)
Misattributed reasoning Correct citation, wrong legal principle Hard (requires substantive analysis)

The third category is most insidious. The citation checks out. The case name is real. But the legal proposition attributed to it isn't actually supported by the decision. This type requires substantive legal analysis to catch—exactly what lawyers were hoping AI would handle.

Case Study: Ko v. Li (Canada)

The Ontario Superior Court's handling of Ko v. Li, 2025 ONSC 2766 established a template for how courts address AI hallucinations.

What happened: A Toronto lawyer submitted a factum in matrimonial proceedings containing case citations that could not be found on CanLII, Westlaw, Quicklaw, or Google. When asked if AI like ChatGPT supported preparation, she said she would need to ask her clerk.

The admission: "The factum I filed contained citations that were inaccurate and, as I now understand, referred to fictitious cases. I deeply regret this and fully accept responsibility. These authorities were drafted using a legal research tool powered by artificial intelligence. While this tool generated legal arguments, I made the grave mistake of failing to verify the case law."

The outcome: The court initiated contempt proceedings, ultimately dismissing them on conditions that included completing Continuing Professional Development courses and not billing the client for the tainted research and factum.

The broader impact: Ontario's Rule 4.06.1(2.1) now requires counsel to certify the authenticity of every authority cited in factums. The rule exists specifically because of AI hallucination risk.

Case Study: Walters v. OpenAI (Georgia)

The May 2025 ruling in Walters v. OpenAI addressed a different question: can AI companies be held liable for hallucinated content that defames real people?

Background: Mark Walters, a nationally syndicated radio host and gun rights advocate, sued OpenAI after ChatGPT incorrectly stated he had been accused of embezzlement in a lawsuit filed by the Second Amendment Foundation. He was never a party to that case.

The court's ruling: Judge Tracie Cason granted summary judgment on three independent grounds:

  1. No defamatory meaning: Given ChatGPT's disclaimers about potential inaccuracies and the journalist's acknowledgment of the error, no reasonable reader would interpret the AI's response as factual assertion.

  2. No negligence or actual malice: OpenAI "leads the AI industry in attempting to reduce and avoid mistaken output." The court rejected the argument that releasing a system known to hallucinate constitutes negligence: "Such a rule would impose a standard of strict liability, not negligence."

  3. No damages: The output was shown only to the editor who prompted it, who never believed or republished it.

The significance: This ruling provides significant protection for AI companies, but it's limited to scenarios where users knew the AI could err and the output wasn't widely published. It doesn't help lawyers who submit hallucinated content to courts.

Pattern Analysis: Who's Getting Caught?

Charlotin's database reveals clear patterns:

Factor Finding
Firm size 90% are solo practitioners or small firms
Party representation 56% plaintiff's counsel; 31% defense counsel
Detection method Usually opposing counsel or court clerk verification
Sanction range $3,000 to $31,000+

The concentration in small firms isn't surprising. Larger firms have developed AI governance policies, mandatory verification workflows, and training programs. Solo practitioners often lack these guardrails.

The MyPillow Case: Recent High-Profile Sanction

In July 2025, Judge Nina Y. Wang of the U.S. District Court in Denver ordered attorneys Christopher Kachouroff and Jennifer DeMaster to pay $3,000 each for filing an AI-generated motion in Mike Lindell's defamation case containing "more than two dozen mistakes—including hallucinated cases."

The attorneys initially claimed the filing was an accidental draft. But the "final" version was still riddled with errors. The court found their "contradictory statements and lack of corroborating evidence" indicated the filing was not inadvertent.

Legal commentators noted the sanction was light. "Recently there was a $31,000 sanction for a generative AI error," noted Above the Law, "which is far more in line with where we are right now."

The message: courts are losing patience. Sanctions are escalating.


Tool Review: AI Legal Research Verification Tools

The emerging category of tools designed to catch AI hallucinations before they reach the court

PelAIkan (Charlotin's Tool)

What It Is: Automated reference checker built from the hallucination database

Best For: Verifying AI-generated legal citations before filing

Key Feature: Trained on the specific failure patterns identified across hundreds of real cases

Status: Research tool; not yet commercial product

Assessment: Proof of concept for what proper verification infrastructure should include


Westlaw AI-Assisted Research

What It Is: Thomson Reuters' RAG-enhanced legal research platform

Best For: Primary legal research with citation verification

Hallucination Rate: Higher than competitors in Stanford testing; 42% accuracy vs. 83%+ for Lexis+ AI

Key Limitation: Despite "AI-assisted" branding, still requires manual verification

Rating: 3/5 for hallucination prevention specifically


Lexis+ AI

What It Is: LexisNexis's generative AI legal research platform

Best For: Research tasks where citations are automatically linked to source documents

Hallucination Rate: ~17% error rate in Stanford study

Key Feature: Direct links to cited materials enable rapid verification

Rating: 4/5 - better performance, but 17% is still too high for blind trust


The Honest Assessment

No AI legal research tool has eliminated hallucinations. The Stanford study found that even purpose-built legal AI tools hallucinate at alarming rates. The solution isn't finding a "hallucination-free" tool—it's building verification into every workflow.


What's Working: Firms That Haven't Been Sanctioned

Success Story: The Verification Protocol

Firm type: Mid-sized US litigation practice

Approach: Mandatory verification workflow for any AI-generated research

Protocol details:

  1. AI output flagged with "UNVERIFIED" watermark
  2. Every citation must be pulled in original source before use
  3. Quotations verified verbatim against reporter text
  4. Second attorney sign-off required on AI-assisted filings

Result: Zero AI-related sanctions; partners report AI still saves significant time despite verification overhead.

Key insight: "We're not anti-AI. We're anti-embarrassment. The verification step adds 20 minutes. Being sanctioned costs 20 hours minimum plus reputation damage."


Success Story: The Training Investment

Firm type: UK regional practice

Challenge: Associates using ChatGPT without understanding limitations

Solution: Mandatory 4-hour training on AI limitations, hallucination examples, and verification procedures

Training includes:

  • Review of real hallucination cases
  • Hands-on exercises identifying fabricated citations
  • Practice using verification tools
  • Clear policy on AI use documentation

Result: Firm-wide AI literacy; zero submissions with hallucinated content

Key insight: "The associates who understand how LLMs work are less likely to blindly trust them. Education is cheaper than sanctions."


Hard Cases: Why Verification Sometimes Fails

Hard Case #1: The Time Pressure Problem

Scenario: Filing deadline in 4 hours; associate uses AI for research to meet it.

What went wrong: No time for comprehensive verification. Citations looked plausible. One was fabricated.

The sanction: $5,000 plus mandatory CLE on AI ethics.

Lesson: AI doesn't eliminate time constraints. If anything, it creates false confidence that work can be done faster than it actually can.


Hard Case #2: The Confident AI

Scenario: Lawyer asks AI to verify its own citations. AI confirms they're accurate.

What went wrong: LLMs generate confident-sounding verification even for fabricated content. The lawyer treated AI self-verification as genuine confirmation.

The court's view: "The attorney cannot outsource professional responsibility to a chatbot."

Lesson: AI verifying its own output is not verification. Independent source confirmation is required.


Hard Case #3: The "Real Case, Wrong Holding" Problem

Scenario: Citation exists. Case name is real. But the legal principle attributed to it isn't in the decision.

What went wrong: Lawyer checked case existed but didn't read the actual holding closely.

Detection: Opposing counsel read the case and filed sanctions motion.

Lesson: Verification means reading the source, not just confirming it exists.


Reliability Corner

The Charlotin Database by Numbers (October 2025)

Metric Count
Total global cases 486+
United States 324
Growth rate 2-3 new cases per day
Typical sanction range $3,000-$31,000+
Cases involving contempt proceedings ~15%

Stanford Hallucination Study Findings

Tool Hallucination Rate
General-purpose LLMs (GPT-4, Claude) 58-88%
Lexis+ AI ~17%
Ask Practical Law AI ~17%
Westlaw AI-Assisted Research 58% (42% accuracy)

Source: Magesh et al., Journal of Empirical Legal Studies, 2025

This Month's Perspective

The gap between "AI legal research" marketing and actual reliability remains significant. Tools marketed as reducing hallucinations still produce them at rates that would be unacceptable in any other professional context.


Workflow of the Month: Pre-Filing Hallucination Prevention Checklist

Use this before submitting any filing that involved AI assistance in research or drafting.

PRE-FILING HALLUCINATION PREVENTION
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━

MATTER: _________________________________
FILING: _________________________________
AI TOOLS USED: __________________________
DATE: __________________________________

CITATION VERIFICATION
□ List all case citations in the document
  (Use separate sheet if needed)

For EACH citation, verify:
□ Case exists in official reporter?
  Checked via: Westlaw / Lexis / Other: _____
□ Case name spelled correctly?
□ Citation format accurate (volume, reporter, page)?
□ Year matches reported decision date?
□ Court designation correct?

QUOTATION VERIFICATION
For EACH direct quotation, verify:
□ Quotation appears verbatim in source?
□ Page/paragraph citation accurate?
□ Quotation not taken out of context?
□ Ellipses accurately reflect omissions?

HOLDING VERIFICATION
For EACH legal proposition, verify:
□ The cited case actually supports this point?
□ The holding hasn't been overruled?
□ The case is from binding/persuasive authority?
□ Read the actual opinion (not just headnotes)?

AI USE DOCUMENTATION
□ Document which portions used AI assistance
□ AI-generated content separately verified?
□ Complies with local court AI disclosure rules?
  Local rule: ______________________________

SECONDARY VERIFICATION
□ Second attorney reviewed AI-assisted portions?
  Reviewer: _______________________________
□ Any flagged concerns resolved?

CERTIFICATION
I have personally verified every citation and
quotation in this filing. I understand that AI
tools may produce plausible-sounding but
fabricated legal authorities.

Signature: _______________ Date: __________

VERIFICATION LOG
Citation 1: _______ Verified: □ Yes □ No
Citation 2: _______ Verified: □ Yes □ No
Citation 3: _______ Verified: □ Yes □ No
Citation 4: _______ Verified: □ Yes □ No
Citation 5: _______ Verified: □ Yes □ No
(Continue on separate sheet)

Time investment: 30-90 minutes depending on filing complexity Why it matters: No sanction ever resulted from over-verification


Quick Hits

Sanctions News:

Regulatory Updates:

  • Ontario Rule 4.06.1(2.1) requires certification of case authenticity in factums
  • Multiple US federal courts now require AI use disclosure

Database Resource:

Coming Next Issue:

  • Contract Intelligence Platforms: Beyond Review to Analytics

Ask the Community

The hallucination problem raises questions we're researching:

  1. For practitioners: Have you implemented AI verification workflows? What's working?
  2. For firms with policies: How do you balance AI efficiency with verification overhead?
  3. For those who've seen sanctions: What could have prevented the issue?
  4. Would you use a standardized pre-filing verification checklist?

Reply to share. Anonymized contributions welcome.


TwinLadder Weekly | Issue #16 | September 2025

Helping lawyers build AI capability through honest education.


Sources