Back to the MIT repository
4. Malicious Actors & Misuse2 - Post-deployment

Defamation / libel / slander

Use of a technology system to create, facilitate or amplify false perception(s) about an individual, group or organisation

Source: MIT AI Risk Repositorymit1353

ENTITY

1 - Human

INTENT

1 - Intentional

TIMING

2 - Post-deployment

Risk ID

mit1353

Domain lineage

4. Malicious Actors & Misuse

223 mapped risks

4.3 > Fraud, scams, and targeted manipulation

Mitigation strategy

1. Implement Security by Design and Pre-deployment Audits: Integrate Adversarial AI (AAI) risk analysis and mitigation early in the system lifecycle, including pre-deployment red-teaming and model audits, to proactively identify and avoid high-risk outputs (e.g., those that generate personal profiles or synthesize information about individuals and organizations). 2. Establish a Multi-Layered Rapid-Response Protocol: Develop and enforce stringent notice-and-takedown protocols with clear escalation triggers across legal, communications, and operations teams to ensure rapid investigation, documentation, and correction or retraction of false, defamatory outputs post-deployment. 3. Mandate Provenance Tracking and Transparent Governance: Institute robust provenance tracking mechanisms, logging all user prompts, system outputs, and subsequent moderation actions to establish a detailed audit trail, which serves to assess fault, mitigate punitive exposure, and enhance model transparency for both internal governance and potential judicial discovery.