Back to the MIT repository
7. AI System Safety, Failures, & Limitations3 - Other

Over- and underfitting

Over- and underfitting describe the over or insufficient adaption of a model to training data. Both phenomena can cause an AI system to behave unreliably if confronted with operational data.

Source: MIT AI Risk Repositorymit1009

ENTITY

3 - Other

INTENT

3 - Other

TIMING

3 - Other

Risk ID

mit1009

Domain lineage

7. AI System Safety, Failures, & Limitations

375 mapped risks

7.3 > Lack of capability or robustness

Mitigation strategy

1. Implement strategic regularization techniques (e.g., L1, L2, Dropout) and dynamically manage model complexity (e.g., varying layer count or feature set size) to navigate the bias-variance trade-off, thereby constraining the capacity to memorize training noise (overfitting) while ensuring sufficient capacity to capture underlying data patterns (underfitting). 2. Employ k-fold cross-validation during the training process in conjunction with early stopping mechanisms. This ensures the model's performance on independent data subsets is robustly evaluated, halting training at the optimal point before generalization performance on the validation set begins to degrade. 3. Maximize the diversity and volume of the training dataset through data augmentation techniques or the collection of additional labeled data, and perform rigorous feature engineering and data cleaning to minimize the impact of noise and outliers that contribute to poor model fit.