Obscene, Degrading, and/or Abusive Content
Eased production of and access to obscene, degrading, and/or abusive imagery which can cause harm, including synthetic child sexual abuse material (CSAM), and nonconsensual intimate images (NCII) of adults.
ENTITY
1 - Human
INTENT
1 - Intentional
TIMING
2 - Post-deployment
Risk ID
mit766
Domain lineage
1. Discrimination & Toxicity
1.2 > Exposure to toxic content
Mitigation strategy
1. Implement stringent safety guardrails and filtering mechanisms during the training and development of generative AI models, coupled with robust pre-deployment safety testing by designated bodies, to prevent the system from acquiring or being exploited to generate Child Sexual Abuse Material (CSAM) or Non-Consensual Intimate Images (NCII) at the source. 2. Mandate the adoption of Digital Content Transparency (DCT) techniques, such as provenance data tracking and digital watermarking, to enable the efficient identification and segregation of synthetic content from non-synthetic content, thereby enabling faster detection and removal of harmful outputs post-deployment. 3. Establish clear regulatory and legal frameworks that impose platform-level accountability for the dissemination of AI-generated CSAM and NCII, requiring technology providers to deploy adequate moderation tools, reporting mechanisms, and to enforce user policies that treat synthetic illegal content no differently than traditional illegal content.