Study Highlights Persistent Hallucinations in Legal AI Systems
May 23, 2024
Stanford University’s Human-Centered AI Institute (HAI) conducted a study in which they designed a "pre-registered dataset of over 200 open-ended legal queries" to test AI products by LexisNexis (creator of Lexis+ AI) and Thomson Reuters (creator of Westlaw AI-Assisted Research and Ask Practical Law AI). The researchers found that these legal models hallucinate in 1 out of 6 (or more) benchmarking queries.
- Alleged deployer
- legal-professionals, law-firms, organizations-requiring-legal-research
- Alleged developer
- thomson-reuters, lexisnexis
- Alleged harmed parties
- legal-professionals, clients-of-lawyers, legal-system
Source
Data from the AI Incident Database (AIID). Cite this incident: https://incidentdatabase.ai/cite/704
Data source
Incident data is from the AI Incident Database (AIID).
When citing the database as a whole, please use:
McGregor, S. (2021) Preventing Repeated Real World AI Failures by Cataloging Incidents: The AI Incident Database. In Proceedings of the Thirty-Third Annual Conference on Innovative Applications of Artificial Intelligence (IAAI-21). Virtual Conference.
Pre-print on arXiv · Database snapshots & citation guide
We use weekly snapshots of the AIID for stable reference. For the official suggested citation of a specific incident, use the “Cite this incident” link on each incident page.