Alleged Fake Citations Undermine Expert Testimony in Minnesota Deepfake Law Case

November 1, 2024

In a recent legal case defending Minnesota's deepfake election misinformation law, Stanford misinformation expert Professor Jeff Hancock's affidavit allegedly cited non-existent academic sources. The reportedly fabricated citations appear to have undermined the credibility of his testimony, raising concerns about AI-generated disinformation and the importance of trustworthy AI governance.

For those interested in shaping safe and secure AI practices, explore HISPI Project Cerebellum TAIM (Govern) for harm prevention through guardrails for AI.
JOIN US

Matched TAIM controls

Suggested mapping from embedding similarity (not a formal assessment). Browse all TAIM controls

Alleged deployer
jeff-hancock
Alleged developer
openai, chatgpt
Alleged harmed parties
mary-franson, keith-ellison, jeff-hancock, christopher-kohls, chad-larson

Source

Data from the AI Incident Database (AIID). Cite this incident: https://incidentdatabase.ai/cite/852

Data source

Incident data is from the AI Incident Database (AIID).

When citing the database as a whole, please use:

McGregor, S. (2021) Preventing Repeated Real World AI Failures by Cataloging Incidents: The AI Incident Database. In Proceedings of the Thirty-Third Annual Conference on Innovative Applications of Artificial Intelligence (IAAI-21). Virtual Conference.

Pre-print on arXiv · Database snapshots & citation guide

We use weekly snapshots of the AIID for stable reference. For the official suggested citation of a specific incident, use the “Cite this incident” link on each incident page.