Back to the MIT repository
4. Malicious Actors & Misuse2 - Post-deployment

Harm to individuals through fake content

Malicious actors can use general- purpose AI to generate fake content that harms individuals in a targeted way. For example, they can use such fake content for scams, extortion, psychological manipulation, generation of non- consensual intimate imagery (NCII) and child sexual abuse material (CSAM), or targeted sabotage of individuals and organisations.

Source: MIT AI Risk Repositorymit1020

ENTITY

1 - Human

INTENT

1 - Intentional

TIMING

2 - Post-deployment

Risk ID

mit1020

Domain lineage

4. Malicious Actors & Misuse

223 mapped risks

4.3 > Fraud, scams, and targeted manipulation

Mitigation strategy

1. Mandatory Technical Safeguards and Content Provenance at Creation Implement rigorous semantic guardrails, including multimodal safety classifiers and data filtering, on both input prompts and model outputs to actively prevent the generation of non-consensual intimate imagery (NCII) and child sexual abuse material (CSAM). Concurrently, integrate the Coalition for Content Provenance and Authentication (C2PA) standards, applying imperceptible watermarking and metadata embedding at the point of creation to establish verifiable content authenticity and provenance. 2. Advanced, Real-Time Deepfake and Misinformation Detection Systems Deploy robust Digital Content Transparency (DCT) and deepfake detection solutions, leveraging advanced deep learning architectures (e.g., CNNs, RNNs) and forensic analysis to identify subtle, non-human-perceptible inconsistencies and artifacts in synthetic media. This detection must be integrated into real-time systems to counter fraud and impersonation attempts during critical processes like identity verification. 3. Systemic Governance and Organizational Resilience Establish a comprehensive AI governance framework rooted in the principles of fairness, transparency, accountability, and security to guide the responsible deployment of AI systems. Furthermore, organizations that rely on or deploy generative AI must prepare for targeted misinformation and fraud attacks by creating and exercising detailed response plans aligned with established crisis communication strategies.