Back to the MIT repository
7. AI System Safety, Failures, & Limitations2 - Post-deployment

Inappropriate degree of automation

The AI application’s degree of automation ranges from no automation to fully autonomous. AI applications with a high degree of automation may exhibit unexpected behaviour and pose risks in terms of their reliability and safety.

Source: MIT AI Risk Repositorymit995

ENTITY

2 - AI

INTENT

2 - Unintentional

TIMING

2 - Post-deployment

Risk ID

mit995

Domain lineage

7. AI System Safety, Failures, & Limitations

375 mapped risks

7.2 > AI possessing dangerous capabilities

Mitigation strategy

1. Mandate Human Oversight and Intervention Mechanisms Implement design requirements that integrate human-in-the-loop decision-making and supervision, proportionate to the AI's degree of autonomy and the criticality of its function. Crucially, engineer and rigorously test fail-safe mechanisms, such as an override or emergency shutdown protocol, to immediately disable or assume control of the autonomous system upon detection of unpredictable or unsafe emergent behavior. 2. Deploy Continuous, Real-Time Operational Monitoring Establish continuous monitoring pipelines that track the AI system's real-time performance, behavioral metrics, and output fidelity. This includes deploying advanced anomaly detection capabilities to promptly identify deviations from intended operational parameters, enabling automated alerting and rapid response to unexpected behavior or potential safety failures post-deployment. 3. Conduct Rigorous Robustness and Resilience Testing Execute comprehensive pre-deployment and ongoing adversarial robustness testing (AI Red Teaming) and stress testing to proactively challenge the system's performance under edge cases, unexpected data inputs, and high-pressure scenarios. This process is essential to minimize latent vulnerabilities and enhance the model's resilience against conditions that could induce unreliable or unsafe operation.