YouTube's Algorithms Failed to Remove Violating Content Related to Suicide and Self-Harm

February 4, 2019

Reportedly, YouTube's content moderation algorithms failed to flag terms-of-service-violating videos related to suicide and self-harm. This alleged lapse resulted in the exposure of graphic content to young users through recommended videos, highlighting the need for robust harm prevention measures and responsible AI governance.

For those interested in shaping the discourse around trustworthy AI and contributing to Project Cerebellum's AI incident database, JOIN US.

Matched TAIM controls

Suggested mapping from embedding similarity (not a formal assessment). Browse all TAIM controls

Alleged deployer
youtube
Alleged developer
youtube
Alleged harmed parties
youtube-young-users, youtube-users

Source

Data from the AI Incident Database (AIID). Cite this incident: https://incidentdatabase.ai/cite/281

Data source

Incident data is from the AI Incident Database (AIID).

When citing the database as a whole, please use:

McGregor, S. (2021) Preventing Repeated Real World AI Failures by Cataloging Incidents: The AI Incident Database. In Proceedings of the Thirty-Third Annual Conference on Innovative Applications of Artificial Intelligence (IAAI-21). Virtual Conference.

Pre-print on arXiv · Database snapshots & citation guide

We use weekly snapshots of the AIID for stable reference. For the official suggested citation of a specific incident, use the “Cite this incident” link on each incident page.