Data bias
Historical and societal biases that are present in the data are used to train and fine-tune the model.
ENTITY
1 - Human
INTENT
2 - Unintentional
TIMING
2 - Post-deployment
Risk ID
mit1279
Domain lineage
1. Discrimination & Toxicity
1.1 > Unfair discrimination and misrepresentation
Mitigation strategy
1. Proactively source and curate diverse and representative training data from multiple, high-quality sources, prioritizing the identification and correction of under-representation or differential data quality issues across all protected groups. 2. Employ pre-processing techniques such as data reweighting, up/down sampling, or synthetic data generation (e.g., SMOTE, GANs) to modify the distribution of training instances, thereby ensuring greater balance and reducing the disproportionate influence of historically biased data patterns. 3. Utilize fair representation learning algorithms to transform the training data into a latent representation that minimizes the encoding of information related to sensitive or protected attributes, thus mitigating the transfer of historical bias into the trained model. 4. Implement rigorous data governance and auditing protocols, including feature blinding (masking or filtering sensitive attributes) during model training, to proactively monitor and restrict the model's reliance on features that serve as proxies for systemic bias.