Back to the MIT repository
7. AI System Safety, Failures, & Limitations2 - Post-deployment

Harms to non-humans

Large-scale harms to animals and the development of AI capable of suffering.

Source: MIT AI Risk Repositorymit1039

ENTITY

1 - Human

INTENT

2 - Unintentional

TIMING

2 - Post-deployment

Risk ID

mit1039

Domain lineage

7. AI System Safety, Failures, & Limitations

375 mapped risks

7.5 > AI welfare and rights

Mitigation strategy

1. Implement rigorous systemic risk assessment and mitigation frameworks, including state-of-the-art model evaluations and enhanced cybersecurity protections for general-purpose AI models, to prevent unforeseen, large-scale negative externalities, particularly those related to issues of control and emergent capabilities such as artificial suffering. 2. Mandate and adhere to energy-efficient and sustainable practices across the AI lifecycle, such as prioritizing renewable-energy-powered data centers, simplifying model architectures, and leveraging transfer learning to minimize the significant carbon and water footprints that contribute to ecological and non-human harm. 3. Establish comprehensive and accessible audit trails and logging mechanisms throughout the AI system's design, development, and deployment phases, ensuring all model behaviors and human decisions are meticulously recorded to enable ex-post-facto analysis and facilitate the timely tracing and remediation of novel harmful outputs.