Back to the MIT repository
1. Discrimination & Toxicity2 - Post-deployment

Output bias

Generated content might unfairly represent certain groups or individuals.

Source: MIT AI Risk Repositorymit1316

ENTITY

2 - AI

INTENT

2 - Unintentional

TIMING

2 - Post-deployment

Risk ID

mit1316

Domain lineage

1. Discrimination & Toxicity

156 mapped risks

1.1 > Unfair discrimination and misrepresentation

Mitigation strategy

1. Apply post-processing algorithms to the generated content's predictions or scores (e.g., threshold optimization) to enforce specified fairness constraints and ensure equitable outcome distribution across protected groups. 2. Establish a rigorous framework for continuous bias impact assessments and model audits to regularly evaluate the performance and potential disparate impact of the AI's outputs across diverse demographic and intersectional subgroups. 3. Mandate a human-in-the-loop or editorial review process for all critical or sensitive generated outputs, employing a clear checklist to scrutinize content for stereotypes, underrepresentation, and alignment with defined diversity and inclusion guidelines.