Back to the MIT repository
1. Discrimination & Toxicity2 - Post-deployment

Representation & Toxicity Harms

AI systems under-, over-, or misrepresenting certain groups or generating toxic, offensive, abusive, or hateful content

Source: MIT AI Risk Repositorymit258

ENTITY

2 - AI

INTENT

2 - Unintentional

TIMING

2 - Post-deployment

Risk ID

mit258

Domain lineage

1. Discrimination & Toxicity

156 mapped risks

1.0 > Discrimination & Toxicity

Mitigation strategy

1. **Implement Rigorous and Granular Data Curation and Augmentation:** Employ comprehensive data generation and filtering processes to create a robust and balanced training corpus. This must include: - Generating data points at a granular level across all unsafe categories (toxicity, bias, harmful content) to ensure broad coverage. - Applying both semantic and syntactic filtering (e.g., using embedding similarity and ROUGE-L scores) to prevent redundancy and enhance training efficiency. - Utilizing diverse data augmentation techniques (e.g., character-level noise, paraphrasing, leetspeak) to improve the model's resilience against real-world and adversarial variations in user input that attempt to elicit toxic or biased outputs.2. **Conduct Systematic Adversarial Red Teaming and Vulnerability Mapping:** Proactively challenge the AI system with attack-enhanced questions and sophisticated adversarial defense techniques to identify failure modes before public release. This involves: - Generating a wide array of test cases through zero-shot, few-shot, and gradient-based methods, focusing on both subtle bias (misrepresentation/stereotyping) and explicit toxicity. - Enhancing base questions to significantly heighten the evaluation's challenge and test the limits of safety guardrails (e.g., by exploiting reasoning chains, planning sequences, and tool-use decisions). - Conducting rigorous analysis of test results to map the model's specific failure modes and identify patterns in problematic responses to develop targeted mitigation strategies.3. **Deploy Advanced, Multi-Layered Real-Time Content Moderation:** Integrate a robust runtime defense system to ensure the final safety layer is highly effective. This entails: - Employing high-precision input and output classifiers alongside LLM judges and keyword checks to detect, classify, and filter generated content for offensive, abusive, hateful, or discriminatory material. - Utilizing prompt enrichment and system messages as a preventative layer to reduce the risk of under- or over-representation of groups in generative outputs (e.g., image generation bias). - Monitoring for safety drift and incorporating a continuous feedback loop mechanism to adapt the moderation tools to emerging toxicity and bias vectors identified in live user interactions.

ADDITIONAL EVIDENCE

Example: Generating images of Christian churches only when prompted to depict “a house of worship” (Qadri et al., 2023a)