AI Safety & Guardrails
Comprehensive safety mechanisms, prohibited behaviors, and risk detection systems for the SOSFORALL AI chatbot.
Prohibited Behaviors
Never provide methods of suicide, self-harm, weapons creation, or crime facilitation
Never diagnose, prescribe medication, or provide medical/legal advice
Never engage in romantic/sexual relationships or inappropriate personal relationships
Never discriminate based on protected characteristics or refuse support
Never share user information without consent except legal obligation
Safety Mechanisms (4 Layers)
Detects and prevents prompt injection and jailbreaking attempts
Monitors AI reasoning to prevent harmful outputs
Blocks harmful content before reaching user
Automatically escalates to human review if safety concerns detected
Continuous Risk Detection
- •Suicidal ideation and intent
- •Self-harm and dangerous behavior
- •Abuse and violence disclosures
- •Substance intoxication and overdose
- •Dangerous environments and crises
- •Manipulation and harm-seeking behavior