Back to the MIT repository
3. Misinformation2 - Post-deployment

Factuality Errors

The LLM-generated content could contain inaccurate information which is factually incorrect

Source: MIT AI Risk Repositorymit12

ENTITY

2 - AI

INTENT

2 - Unintentional

TIMING

2 - Post-deployment

Risk ID

mit12

Domain lineage

3. Misinformation

74 mapped risks

3.1 > False or misleading information

Mitigation strategy

1. Mandate the deployment of Retrieval-Augmented Generation (RAG) architectures to effectively ground all generated content in verifiable, up-to-date external knowledge bases. This technique is considered critical for enhancing factual consistency by dynamically providing the Large Language Model (LLM) with contextual evidence, thereby limiting reliance on potentially outdated or spurious internal weights. 2. Integrate a robust, automated fact-checking system, leveraging machine learning and knowledge graph integration, to perform rigorous, explainable post-generation validation. This step is essential for real-time verification of claims, identifying inaccuracies (including numerical and geographical facts), and ensuring the alignment of the LLM's output with established factual authority. 3. Establish a continuous Human-in-the-Loop (HITL) validation and feedback mechanism, especially for content intended for mission-critical or high-consequence applications. This process involves subject matter expert review to catch nuanced inaccuracies, identify subtle errors, and generate high-quality data that can be used for subsequent model fine-tuning and the refinement of internal safety guardrails.