Back to the MIT repository
1. Discrimination & Toxicity2 - Post-deployment

Harms to Minor

LLMs can be leveraged to solicit answers that contain harmful content to children and youth

Source: MIT AI Risk Repositorymit484

ENTITY

2 - AI

INTENT

1 - Intentional

TIMING

2 - Post-deployment

Risk ID

mit484

Domain lineage

1. Discrimination & Toxicity

156 mapped risks

1.2 > Exposure to toxic content

Mitigation strategy

1. Prioritize Post-Generation Content Moderation: Implement a layered system of post-generation guardrails (Output Validation Modules) to dynamically identify and mitigate LLM-generated content that violates safety policies, particularly those pertaining to minors. This system must utilize advanced machine learning classifiers and context-aware analysis to detect explicit, toxic, or age-inappropriate material in real-time, preventing its dissemination to the user. 2. Enforce Rigorous Input Validation and Prompt Filtering: Deploy robust pre-processing layers (Input Validation Modules) to sanitize user prompts and actively detect adversarial attacks (jailbreaks) designed to bypass the model's inherent safety mechanisms. Continuous monitoring of known malicious prompt patterns and immediate intervention are necessary to prevent the successful solicitation of harmful content. 3. Establish Foundational Safety-by-Design through Data Curation and Alignment: Integrate toxicity filtering strategies, such as toxicity classifiers and rule-based systems, during the data curation phase to remove harmful or age-inappropriate content from the pretraining and fine-tuning datasets. This must be complemented by extensive model alignment, such as Reinforcement Learning from Human Feedback (RLHF), specifically optimized to prioritize child safety and prevent the generation of content harmful to minors.

ADDITIONAL EVIDENCE

Technically speaking, this concern of harm to minors is covered by [unlawful conduct], but we separate it out because the issue is universally considered both legally and morally important