Back to the MIT repository
6. Socioeconomic and Environmental2 - Post-deployment

Inequality

More broadly, bad decisions or errors by AI tools could lead to discrimination or deeper inequality

Source: MIT AI Risk Repositorymit909

ENTITY

2 - AI

INTENT

2 - Unintentional

TIMING

2 - Post-deployment

Risk ID

mit909

Domain lineage

6. Socioeconomic and Environmental

262 mapped risks

6.2 > Increased inequality and decline in employment quality

Mitigation strategy

- Implement stringent data governance protocols and rigorous data auditing to ensure training datasets are diverse, representative, and balanced across sensitive demographic attributes via pre-processing techniques like reweighting or synthetic data generation. - Apply fairness-aware algorithmic constraints and in-processing techniques (e.g., adversarial debiasing or fair representation learning) during model training to minimize algorithmic bias and promote equitable decision outcomes. - Establish a continuous and robust AI governance framework that mandates post-deployment performance monitoring for bias drift, incorporates Explainable AI (XAI) for transparency, and defines clear human-in-the-loop processes for critical decision oversight.