Back to the MIT repository
3. Misinformation2 - Post-deployment

Confabulation

The production of confidently stated but erroneous or false content (known colloquially as “hallucinations” or “fabrications”) by which users may be misled or deceived.

Source: MIT AI Risk Repositorymit757

ENTITY

2 - AI

INTENT

2 - Unintentional

TIMING

2 - Post-deployment

Risk ID

mit757

Domain lineage

3. Misinformation

74 mapped risks

3.1 > False or misleading information

Mitigation strategy

1. Implement Retrieval-Augmented Generation (RAG) frameworks to systematically ground model outputs in verified, external knowledge repositories. This mechanism supplies the model with factual, up-to-date context at the time of inference, significantly reducing the propensity for probabilistic fabrication. 2. Prioritize the utilization of high-quality, curated, and domain-specific training datasets, coupled with targeted fine-tuning (e.g., Instruction Tuning or Reinforcement Learning from Human Feedback, RLHF) to enhance the model's internal factual knowledge and reduce knowledge gaps that necessitate inventive responses. 3. Establish robust, multi-layered output validation protocols, including inference-time interventions such as confidence scoring, self-refinement techniques, and the deployment of systematic guardrails to detect and prevent the dissemination of ungrounded or speculative content.