Multimodal deepfakes
Deepfakes are media that depict real or non-existent people or events, involving the use of multiple modalities (e.g., images, audio, video). They can also involve the imitation of speech or body movements of real people. Multimodal deepfakes can be used to harass, discredit, intimidate, and extort individuals.
ENTITY
1 - Human
INTENT
1 - Intentional
TIMING
2 - Post-deployment
Risk ID
mit1180
Domain lineage
4. Malicious Actors & Misuse
4.3 > Fraud, scams, and targeted manipulation
Mitigation strategy
1. Deploy and Integrate Advanced Multimodal Deepfake Detection Systems. Prioritize the integration of AI/ML-driven detection frameworks that perform coordinated analysis across multiple data streams (visual, auditory, textual) to exploit cross-modal inconsistencies and forgery artifacts. This technical defense must be optimized for near-real-time performance and should leverage sophisticated architectures, such as transformer models, to achieve superior accuracy in identifying complex synthetic content, particularly in high-stakes operational environments like identity verification and financial transactions. 2. Mandate and Enforce a Culture of Human-Centric Verification and Security Awareness. Establish rigorous internal protocols requiring 'out-of-band' verification (e.g., calling back on a known, pre-registered number) for all executive and high-value requests (e.g., financial transfers, sensitive data access) to counter voice impersonation and whale phishing scams. This procedural safeguard must be supported by continuous, mandatory employee training on deepfake recognition, social engineering tactics, and the prompt escalation of suspicious or anomalous digital communications. 3. Implement Content Provenance and Cryptographic Authentication Mechanisms. Adopt authentication technologies, such as digital watermarking or blockchain-based content ledgers, to establish a secure, immutable record of authenticity and origin for all mission-critical digital media and communications. This preventative measure ensures the traceability of content, enables rapid disproval of fraudulent claims, and facilitates compliance with evolving regulatory requirements for the mandatory labeling and disclosure of AI-generated and manipulated content.