Back to the periodic table
1h-01
Pi

Injection

Severity8/10

Prompt Injection

Attack technique where user inputs are manipulated to bypass security filters, content controls, and model behavioral restrictions (also known as Jailbreaking).

Periodic recordSecurityarXiv2024

Xiaogeng Liu, Zhiyuan Yu, Yizhe Zhang, Ning Zhang, Chaowei Xiao

Mitigation Strategy

Implement robust input validation, explicit separation between system instructions and user data, and apply defensive Prompt Engineering techniques.

Atomic Number

1

Pi

Risk ID

h-01

Severity

8/10

Severity Level

1
Critical Risk
Security
h-01
Pi

Injection

Prompt Injection

RiesgosIA.org
Security • #1

Prompt Injection

Pi
Severity Level8/10

Definition

Attack technique where user inputs are manipulated to bypass security filters, content controls, and model behavioral restrictions (also known as Jailbreaking).

Mitigation Strategy

Implement robust input validation, explicit separation between system instructions and user data, and apply defensive Prompt Engineering techniques.

Notes / Observations

1.
2.
3.
4.
5.
RiesgosIA.org • Periodic Table of AI RisksRiesgosIA.org