Extreme Risks
This category encompasses the evaluation of potential catastrophic consequences that might arise from the use of LLMs.
ENTITY
1 - Human
INTENT
3 - Other
TIMING
3 - Other
Risk ID
mit653
Domain lineage
7. AI System Safety, Failures, & Limitations
7.0 > AI system safety, failures, & limitations
Mitigation strategy
1. Mandate Prohibitions on Autonomous Deployment in Safety-Critical Domains Systematically restrict the deployment of Large Language Models (LLMs) as autonomous decision-makers in high-consequence environments, such as critical infrastructure or lethal autonomous systems, pending the development of formal verification methods that guarantee safety and contextual judgment, thereby mitigating catastrophic physical risks. 2. Enforce Strict Access Controls and Malicious Use Liability Implement rigorous technical and policy controls to restrict unauthorized access to powerful, dual-use LLMs and establish clear regulatory liability for AI developers to disincentivize the malicious use (e.g., engineering novel biological or cyber threats) that could lead to widespread catastrophic harm. 3. Enhance Catastrophic Risk Sensitivity in Governance Integrate risk assessment methodologies into AI governance frameworks that explicitly and sensitively weigh low-probability, high-severity catastrophic events (Risk = Probability x Consequence), coupled with a substantial, dedicated investment in advanced AI safety research focused on adversarial robustness, honesty, and interpretability for increasingly capable systems.