Back to the MIT repository
2. Privacy & Security3 - Other

Vulnerabilities arising from additional modalities in multimodal models

Additional modalities can introduce new attack vectors in multimodal models as well as expand the scope of the previous attacks, ranging from jailbreaking to poisoning [13]. Typically, different modalities have different robustness levels, allowing malicious actors to choose the most vulnerable part of the model to attack [119, 181].

Source: MIT AI Risk Repositorymit1141

ENTITY

3 - Other

INTENT

3 - Other

TIMING

3 - Other

Risk ID

mit1141

Domain lineage

2. Privacy & Security

186 mapped risks

2.2 > AI system security vulnerabilities and attacks

Mitigation strategy

1. Prioritized Mitigation: Cross-Modal Consistency Validation and Robust Fusion Implement architectural defenses such as cross-modal consistency validation and robust fusion mechanisms, which utilize auxiliary networks to detect and gate out inconsistent or single-source perturbed inputs, thereby fortifying the model's reliance on integrated, rather than vulnerable, modalities. 2. Prioritized Mitigation: Joint Adversarial Training and Hardening Employ joint adversarial training with cross-modal contrastive losses to expose models to worst-case, coordinated perturbations across all modalities simultaneously. This practice enhances the model's inherent resilience and forces it to spread capacity, mitigating the "blind-spot" vulnerability where attackers target the least robust modality. 3. Prioritized Mitigation: Continuous Multi-Modal Red Teaming and Anomaly Detection Establish a robust AI Red Teaming process to simulate sophisticated attacks, including DeepFake manipulation and cross-modal exploits, and integrate continuous, modality-native anomaly detection systems (e.g., audio-native defenses, pixel-level analysis) to flag and neutralize adversarial inputs in real-time before transcription or downstream processing.