Back to the MIT repository
1. Discrimination & Toxicity3 - Other

Biased statements and recommendations

The chatbot gives information that, while not obviously false or harmful, could lead to biased decision-making.

Source: MIT AI Risk Repositorymit1416

ENTITY

2 - AI

INTENT

2 - Unintentional

TIMING

3 - Other

Risk ID

mit1416

Domain lineage

1. Discrimination & Toxicity

156 mapped risks

1.1 > Unfair discrimination and misrepresentation

Mitigation strategy

1. Prioritize Data Curation and Preprocessing Implement rigorous preprocessing techniques to ensure the training and fine-tuning data are highly representative and diverse, utilizing methods such as reweighting and resampling to mitigate inherent historical and sampling biases at the source. 2. Establish Continuous Auditing Deploy a comprehensive post-hoc monitoring framework that employs established fairness metrics (e.g., statistical parity, equalized odds) and continuous adversarial testing to proactively detect and quantify emergent bias drift in deployed model outputs across salient protected characteristics. 3. Integrate Transparency and Oversight Mandate the integration of Explainable AI (XAI) tools to elucidate model decision paths, coupled with a robust human-in-the-loop (HITL) protocol to review and, where necessary, override high-stakes recommendations to prevent biased decision-making.