Model sensitivity to prompt formatting
LLMs can be highly sensitive to variations in prompt formatting, such as changes in separators, casing, or spacing. Even minor modifications can lead to significant shifts in model performance, potentially affecting the reliability of model evaluations and comparisons. This sensitivity persists across different model sizes and few-shot examples [177].
ENTITY
2 - AI
INTENT
3 - Other
TIMING
2 - Post-deployment
Risk ID
mit1146
Domain lineage
7. AI System Safety, Failures, & Limitations
7.3 > Lack of capability or robustness
Mitigation strategy
1. Implement Comprehensive Multi-Prompt Robustness Evaluation The organization must establish a protocol for quantifying "prompt brittleness" by systematically evaluating model performance across an ensemble of semantically equivalent prompt formats (e.g., JSON, YAML, Markdown, plain text). For high-risk applications, employ prompt ensembling techniques—such as the Mixture of Formats (MOF) approach—to diversify prompt styles within few-shot examples, thereby mitigating the model's over-reliance on any single format's idiosyncratic stylistic features and stabilizing overall reliability. 2. Enforce Strict Prompt Formatting and Structure Standardization Mandate the use of clear, explicit formatting conventions and structural constraints within all production prompts. This includes utilizing dedicated system messages, unambiguous separators (e.g., "\#\#\#" or XML tags) to delineate instructions, context, and data, and specifying output schemas (e.g., JSON format) to guide the model. Such engineering discipline reduces ambiguity and provides the model with consistent structural signals, preempting performance volatility associated with minor variations in casing, spacing, or punctuation. 3. Establish Prompt Consistency and Change Control Treat high-value prompts as critical configuration assets subject to formal change management. Implement consistency checks and version control to ensure that core prompts, especially those used for system alignment or safety guardrails, remain functionally identical across model updates and continuous deployment cycles. Any necessary prompt modification must be accompanied by a dedicated re-evaluation against the established multi-prompt robustness test suite to detect and correct any unintended behavioral drift prior to deployment.