Facial Recognition in Remote Learning Software Reportedly Failed to Recognize a Black Student’s Face

February 1, 2021

A student, who is African American, encountered an issue with remote-proctoring software during a lab quiz check-in. The software failed to recognize her face, leading to excessive changes of her environment to ensure proper functioning.

This incident highlights the need for safe and secure AI practices in education. Such incidents can be mapped to the HISPI Project Cerebellum TAIM (Measure) function to identify areas for improvement in guardrails for AI and promote trustworthy AI.

For those interested in shaping responsible AI governance, JOIN US.

Matched TAIM controls

Suggested mapping from embedding similarity (not a formal assessment). Browse all TAIM controls

Alleged deployer
unknown
Alleged developer
unknown
Alleged harmed parties
amaya-ross, black-students, black-test-takers

Source

Data from the AI Incident Database (AIID). Cite this incident: https://incidentdatabase.ai/cite/158

Data source

Incident data is from the AI Incident Database (AIID).

When citing the database as a whole, please use:

McGregor, S. (2021) Preventing Repeated Real World AI Failures by Cataloging Incidents: The AI Incident Database. In Proceedings of the Thirty-Third Annual Conference on Innovative Applications of Artificial Intelligence (IAAI-21). Virtual Conference.

Pre-print on arXiv · Database snapshots & citation guide

We use weekly snapshots of the AIID for stable reference. For the official suggested citation of a specific incident, use the “Cite this incident” link on each incident page.