AI as Jury: UNC Law School Tests ChatGPT, Grok & Claude in a Mock Trial

No Comments

University of North Carolina School of Law staged an unusual courtroom experiment: an entire “jury” made up of AI systems — ChatGPT (OpenAI), Grok (xAI), and Claude (Anthropic). Three tall displays stood in for jurors as the bots “listened” to a fictional robbery case to probe how AI might influence legal decision-making.

Why this matters

AI tools are already common in legal work for research and drafting. Adoption is rising and many firms report ROI — yet failures persist, from fabricated caselaw to misread facts and hidden biases. Several attorneys have even faced fines for filing AI-generated briefs with bogus citations.

What went wrong (for now)

  • AI can’t read body language or tone in the courtroom.
  • Lacks lived human experience and empathy in judgment.
  • Can badly misinterpret text due to simple typos.
  • Risk of racial and contextual bias creeping into “verdicts.”

Post-trial panelists were broadly critical; many attendees left thinking that “trial-by-bot” isn’t a good idea yet.

Technology will recursively repair its way into every human space if we let it — including the jury box.

Could it improve?

Proponents argue these gaps may narrow: video feeds for non-verbal cues, richer context to reduce errors, and steady model upgrades. The harder question is how far AI should be allowed into the justice system — and who is accountable when it errs.

Leave a Reply

Your email address will not be published. Required fields are marked *

Fill out this field
Fill out this field
Please enter a valid email address.
You need to agree with the terms to proceed

Post Author

Previous Post
AI Has Written Half the Internet. Now What?
Next Post
How Bilingual Brains Handle Difficult Tasks More Efficiently
You might also like