Back to the MIT repository
4. Malicious Actors & Misuse2 - Post-deployment

Non-consensual intimate imagery (NCII)

Create sexual explicit material using an adult person’s likeness

Source: MIT AI Risk Repositorymit1253

ENTITY

1 - Human

INTENT

1 - Intentional

TIMING

2 - Post-deployment

Risk ID

mit1253

Domain lineage

4. Malicious Actors & Misuse

223 mapped risks

4.3 > Fraud, scams, and targeted manipulation

Mitigation strategy

1. Prioritized Mitigation: Implement rigorous pre-deployment and continuous runtime 'red teaming' and safety testing protocols to proactively identify and neutralize model vulnerabilities that could allow the generation of non-consensual intimate imagery (NCII). Enforce strict input/output content filtering at the system level to block and log prompts violating terms of service regarding the sexual exploitation of human likeness. 2. Content Provenance and Distribution Interventions: Integrate and enforce the use of content hashing technology (e.g., digital fingerprinting) across all content distribution channels to enable the rapid detection, blocking, and removal of known NCII upon subsequent attempted sharing, effectively curtailing scaled distribution. 3. Organizational and Regulatory Alignment: Establish clear, internal AI governance policies that explicitly prohibit the use of generative AI capabilities for the creation of NCII, ensuring full compliance with evolving state and international legislation that criminalizes the production and distribution of non-consensual deepfakes.