Output bias
Generated content might unfairly represent certain groups or individuals.
ENTITY
2 - AI
INTENT
2 - Unintentional
TIMING
2 - Post-deployment
Risk ID
mit1316
Domain lineage
1. Discrimination & Toxicity
1.1 > Unfair discrimination and misrepresentation
Mitigation strategy
1. Apply post-processing algorithms to the generated content's predictions or scores (e.g., threshold optimization) to enforce specified fairness constraints and ensure equitable outcome distribution across protected groups. 2. Establish a rigorous framework for continuous bias impact assessments and model audits to regularly evaluate the performance and potential disparate impact of the AI's outputs across diverse demographic and intersectional subgroups. 3. Mandate a human-in-the-loop or editorial review process for all critical or sensitive generated outputs, employing a clear checklist to scrutinize content for stereotypes, underrepresentation, and alignment with defined diversity and inclusion guidelines.