Back to the MIT repository
6. Socioeconomic and Environmental2 - Post-deployment

Copyright Violation

LLM systems may output content similar to existing works, infringing on copyright owners.

Source: MIT AI Risk Repositorymit16

ENTITY

1 - Human

INTENT

1 - Intentional

TIMING

2 - Post-deployment

Risk ID

mit16

Domain lineage

6. Socioeconomic and Environmental

262 mapped risks

6.3 > Economic and cultural devaluation of human effort

Mitigation strategy

1. Implement robust, real-time input and output filtering guardrails. Deploy automated checks, such as embedding similarity and cryptographic hashing against blocklists of known copyrighted works, to intervene during output generation and prevent the release of infringing text, code, or creative works. This includes preemptively blocking or redirecting user prompts that explicitly request protected content. 2. Apply machine unlearning and advanced alignment techniques to regulate model behavior. Systematically employ machine unlearning methods to selectively remove or diminish the memorized knowledge of specific copyrighted material from the LLM's trained parameters. Concurrently, utilize efficient fine-tuning and alignment approaches to instill a policy-driven reluctance in the model to generate content matching protected works, ensuring responses default to disclaimers or refusals for infringing queries. 3. Establish a rigorous pre-training data governance and transparency framework. Conduct comprehensive legal and technical vetting of all data sources used for training to minimize the risk of incorporating illicitly acquired or unlawfully used material. Additionally, implement transparency measures, such as standardized dataset documentation and influence analysis, to facilitate traceability of output to specific training data, aiding in both risk assessment and compliance with emerging regulatory requirements.