Back to the MIT repository
1. Discrimination & Toxicity2 - Post-deployment

Discrimination, toxicity, and bias

AI models and the tools that use them may exacerbate unequal access to employment and services. AI-generated content can promote inequality and harmful stereotypes.

Source: MIT AI Risk Repositorymit876

ENTITY

2 - AI

INTENT

2 - Unintentional

TIMING

2 - Post-deployment

Risk ID

mit876

Domain lineage

1. Discrimination & Toxicity

156 mapped risks

1.1 > Unfair discrimination and misrepresentation

Mitigation strategy

1. Implement rigorous, continuous bias audits and data governance, spanning from training data selection (to ensure diversity and representation) through post-deployment monitoring, actively remediating systemic bias encoded in proxy variables and flawed feature selection. 2. Establish mandatory human-in-the-loop oversight and clear accountability frameworks, particularly for high-risk applications such as employment and service access, enabling human review and override of potentially discriminatory algorithmic decisions. 3. Advance algorithmic transparency and explainability by adopting interpretable models and deploying Explainable AI (XAI) techniques, thereby allowing for the systematic analysis and articulation of a model's decision-making process concerning equitable outcomes.