Back to the MIT repository
7. AI System Safety, Failures, & Limitations3 - Other

First-Order Risks

First-order risks can be generally broken down into risks arising from intended and unintended use, system design and implementation choices, and properties of the chosen dataset and learning components.

Source: MIT AI Risk Repositorymit187

ENTITY

3 - Other

INTENT

3 - Other

TIMING

3 - Other

Risk ID

mit187

Domain lineage

7. AI System Safety, Failures, & Limitations

375 mapped risks

7.0 > AI system safety, failures, & limitations

Mitigation strategy

1. Implement continuous data validation and integrity pipelines across the entire AI lifecycle, encompassing periodic auditing of training datasets for bias, representativeness, and drift. This proactive measure is essential to ensure foundational model alignment and mitigate the risk of skewed or discriminatory system outcomes. 2. Employ comprehensive adversarial stress-testing, often referred to as AI red teaming, prior to deployment. This rigorous exercise should be designed to expose vulnerabilities, edge cases, and potential for input/output manipulation, subsequently informing the integration of adversarial training methods to enhance model robustness and resilience. 3. Establish a formal, continuous AI Risk Management Framework, such as the NIST AI RMF, to integrate governance, measurement, and monitoring of system performance and safety. This includes assigning clear accountability to risk owners and designing human-in-the-loop mechanisms for high-stakes decisions and unexpected model behavior.