Legal AI Safety

Ark vs ChatGPT for Legal Research: Why Citation Verification Matters

By Ark Legal AI ~10 min read

In 2023, a New York attorney submitted a brief citing six cases that didn't exist. ChatGPT made them up. The judge sanctioned him and his firm. Three years later, attorneys are still getting disciplined for citing AI hallucinations. Here's why ChatGPT can't safely do legal research — and how Ark solves it.

The Mata v. Avianca Story

Roberto Mata sued Avianca Airlines after a metal serving cart struck his knee on a flight. His attorney, Steven Schwartz of Levidow, Levidow & Oberman, filed a brief opposing Avianca's motion to dismiss. The brief cited six federal cases, including Varghese v. China Southern Airlines, Shaboon v. Egypt Air, and Miller v. United Airlines.

Avianca's lawyers tried to find the cases. They couldn't. Neither could the judge. They didn't exist.

Schwartz had asked ChatGPT to do the research. ChatGPT confidently produced quotations, case names, citations, and even fake docket numbers. Schwartz copy-pasted them into his brief without verifying.

Mata v. Avianca, Inc., 22-cv-1461 (S.D.N.Y. June 22, 2023)

Judge P. Kevin Castel sanctioned Schwartz and his firm $5,000 and required them to notify each judge falsely identified as the author of the fake opinions. The case became the defining cautionary tale for lawyers using generative AI.

The judge's opinion pulled no punches: Schwartz had "abandoned their responsibilities" by "submitting non-existent judicial opinions with fake quotes and citations created by the artificial intelligence tool ChatGPT." The opinion has been cited in dozens of subsequent sanctions orders.

Why ChatGPT Hallucinates Legal Citations

ChatGPT is a general-purpose language model. It predicts the next most likely word based on patterns from its training data. When you ask it for a case citation, it doesn't search a legal database — it generates text that looks like a case citation.

A fake citation like Varghese v. China Southern Airlines, 925 F.3d 1339 (11th Cir. 2019) has all the features of a real citation:

  • Plaintiff and defendant names (sound plausible for aviation litigation)
  • Volume number and reporter (925 F.3d is a real volume)
  • Page number
  • Court and year

It looks exactly like a real case. It just isn't one. ChatGPT generated it the same way it generates any other fluent English — by pattern-matching on what citations usually look like.

Even in 2026, after multiple model upgrades, ChatGPT continues to fabricate citations. OpenAI has added warnings, but the underlying architecture doesn't retrieve from a verified legal database. It generates.

How Ark Prevents Hallucinations

Ark is built on a different architecture: retrieval-augmented generation with strict citation binding. Here's the pipeline:

  1. Your question is converted to a semantic vector using Voyage embeddings.
  2. That vector searches the Ark database: 7.4M case opinions, 26.6M salient points, 142K court rules, all 50-state statutes.
  3. Isaacus Kanon-2, a legal-specialized reranker, reorders the top candidates by relevance.
  4. The top-ranked real cases are passed to Claude Opus 4.7 as context.
  5. The AI generates an answer citing only the cases in the retrieved context.
  6. Every citation is a clickable link to the actual case text in Ark's database.

Because the AI can only cite cases that exist in the retrieval step, it cannot fabricate them. If Ark doesn't find a relevant case, it tells you. It doesn't make one up.

ChatGPT vs Ark: Side-by-Side

FeatureArkChatGPT
Legal database7.4M cases, 26.6M salient pointsNone (training data only)
Citation verificationEvery citation grounded in retrieved caseUnverified — hallucinations common
Clickable citationsYes — link to actual case textNo
State-specific searchYes, per-state indexesNo
Court rules coverage142K+ rules, all 50 statesInconsistent, often outdated
Statute coverageAll 50 states + DC, currentCutoff-dependent, not authoritative
Sanctions riskLow — every cite is realHigh — documented case law
Cost for legal use$29/month$20/month (but not safe for filings)

Don't risk sanctions. Use AI legal research that cites real cases, not hallucinations. Try Ark free — no credit card required.

Start Free Trial

Other AI Sanctions Cases (Post-Mata)

Mata wasn't an isolated incident. Since 2023, courts have sanctioned lawyers repeatedly for filing briefs with AI-generated fake citations:

  • Park v. Kim (2nd Cir. 2024): Attorney sanctioned for citing non-existent cases generated by AI. The Second Circuit issued a formal reprimand.
  • Morgan & Morgan (multiple incidents 2024-2025): Several attorneys at the large plaintiffs' firm faced discipline for fabricated AI citations. The firm issued internal policies restricting AI research tools.
  • Numerous state-court sanctions: New York, California, Texas, Florida, and Michigan courts have all issued sanctions for AI-fabricated legal research since 2024.

The pattern is consistent: an attorney uses ChatGPT (or a similar general-purpose LLM), copies citations without verification, opposing counsel catches the fabrication, court sanctions follow. The fix is simple — use a tool that can't hallucinate.

People Also Ask

Can I use ChatGPT for legal research if I verify the citations myself?

Yes, in theory. In practice, verification means looking up every single citation ChatGPT produces — which takes longer than doing the research the traditional way. You're also unlikely to catch every hallucination, because fake citations look exactly like real ones. Tools like Ark that verify citations automatically give you AI speed without the verification burden.

What's the difference between retrieval-augmented AI and regular ChatGPT?

Regular ChatGPT generates text based on patterns in its training data. Retrieval-augmented AI (like Ark) first searches a real database for relevant documents, then generates answers using only those retrieved documents as context. The AI cannot cite what it doesn't retrieve. This eliminates the category of error that caused Mata v. Avianca.

Does Claude or Gemini hallucinate legal citations too?

Yes. Any general-purpose LLM — Claude, Gemini, Llama, ChatGPT — will fabricate legal citations when asked. The fix isn't switching models; it's adding retrieval grounding. Ark uses Claude Opus 4.7 as its language model, but Opus never generates free-form citations — it only cites what Ark's retrieval pipeline provides.

Are courts requiring disclosure of AI-assisted briefs now?

Some judges in the 5th Circuit, Eastern District of Texas, and Northern District of Illinois have issued standing orders requiring disclosure of generative AI use in filings. More jurisdictions are likely to follow. When you use Ark, you can disclose AI assistance and still rely on your citations because every one is a real case.

Is it malpractice to cite a hallucinated case?

Submitting fabricated citations without verification is a serious violation of Rule 11 (federal) and analogous state rules requiring attorneys to certify the accuracy of filings. Multiple bar associations have issued ethics opinions classifying AI-hallucination-based filings as potential malpractice. Using verified tools like Ark substantially reduces this risk.

The Verdict

ChatGPT is a remarkable general-purpose tool. It's not a legal research tool. Using it for case research is like using a stopped clock to set your watch — it looks right, but the accuracy is coincidental.

Use ChatGPT for: brainstorming, drafting non-citation prose, summarizing documents you provide it, rephrasing passages, generating first drafts.

Don't use ChatGPT for: finding cases, finding statutes, finding court rules, generating citations, anything that gets filed with a court.

Use Ark for: AI-powered legal research where every citation is a verified, clickable, real case. $29/month. No hallucinations.

Stop risking sanctions. Ark gives you the speed of AI with the safety of verified citations.

Get Started Free

Related articles: Ark Legal AI homePricing plansArk vs WestlawArk vs Casetext / CoCounsel