HISPI Project Cerebellum
AI Incidents

Study Highlights Persistent Hallucinations in Legal AI Systems

May 23, 2024

Stanford University’s Human-Centered AI Institute (HAI) conducted a study in which they designed a "pre-registered dataset of over 200 open-ended legal queries" to test AI products by LexisNexis (creator of Lexis+ AI) and Thomson Reuters (creator of Westlaw AI-Assisted Research and Ask Practical Law AI). The researchers found that these legal models hallucinate in 1 out of 6 (or more) benchmarking queries.
Alleged deployer
legal-professionals, law-firms, organizations-requiring-legal-research
Alleged developer
thomson-reuters, lexisnexis
Alleged harmed parties
legal-professionals, clients-of-lawyers, legal-system

Source

Data from the AI Incident Database (AIID). Cite this incident: https://incidentdatabase.ai/cite/704

Data source

Incident data is from the AI Incident Database (AIID).

When citing the database as a whole, please use:

McGregor, S. (2021) Preventing Repeated Real World AI Failures by Cataloging Incidents: The AI Incident Database. In Proceedings of the Thirty-Third Annual Conference on Innovative Applications of Artificial Intelligence (IAAI-21). Virtual Conference.

Pre-print on arXiv · Database snapshots & citation guide

We use weekly snapshots of the AIID for stable reference. For the official suggested citation of a specific incident, use the “Cite this incident” link on each incident page.