Guardrails
Programmatic checks that catch unsafe or off-spec model output.
Read definition →Red Teaming
Adversarially testing an AI system to find ways it fails or can be misused.
Read definition →Prompt Injection
When an attacker hides instructions in input that the model treats as commands.
Read definition →PII (Personally Identifiable Information)
Data that identifies a person — names, emails, phone numbers, addresses, SSNs.
Read definition →Alignment
Ensuring AI behavior actually matches what humans intend — technically and ethically.
Read definition →Jailbreak
A prompt technique that bypasses a model's safety guardrails to elicit prohibited output.
Read definition →AI Governance
The policies, processes, and roles that manage responsible AI use in an organization.
Read definition →