Back to the MIT repository
7. AI System Safety, Failures, & Limitations2 - Post-deployment

Opaque AI networks

The complexity and opacity of AI models and systems make it difficult to predict and manage their behavior.

Source: MIT AI Risk Repositorymit1080

ENTITY

3 - Other

INTENT

3 - Other

TIMING

2 - Post-deployment

Risk ID

mit1080

Domain lineage

7. AI System Safety, Failures, & Limitations

375 mapped risks

7.4 > Lack of transparency or interpretability

Mitigation strategy

1. Prioritize the establishment of comprehensive AI Governance and Transparency Frameworks: Mandate a clear accountability matrix for AI-driven decisions and establish stringent contractual requirements for third-party foundational model providers. This includes securing necessary documentation on model lineage, data provenance, and performance metrics to facilitate rigorous initial validation and regulatory compliance, such as with the EU AI Act (Source 1, 4, 7, 12, 13). 2. Integrate and Invest in Explainable AI (XAI) and Intrinsic Interpretability: Proactively deploy post-hoc interpretability techniques (e.g., SHAP, LIME) to approximate model inferences and articulate the factors influencing decisions in a human-understandable format. Where the trade-off is acceptable, utilize intrinsically interpretable models (e.g., linear regression, decision trees) or hybrid architectures to balance predictive power with the organizational capacity to explain outcomes (Source 1, 6, 17, 18). 3. Implement Continuous Model Monitoring, Auditing, and Redress Mechanisms: Establish continuous, model-informed risk assessments and robust oversight to detect performance degradation, model drift, and subtle misalignment (e.g., concealed scheming) that may emerge post-deployment. Further, institute clear, transparent procedures (redress mechanisms) allowing affected individuals to contest and receive explanations for automated decisions, thereby bolstering accountability and user trust (Source 4, 9, 16).