Skip to main content

What are guardrails in Bolna Voice AI?

Guardrails are safety mechanisms that help you control and filter inappropriate content, prevent unwanted conversations, and ensure your Voice AI agents maintain professional boundaries during interactions.
Bolna Voice AI Playground guardrails configuration interface showing how to add safety rules, content filters, and inappropriate language detection for AI voice agents

Guardrails on Bolna playground


How to add guardrails to your agent

Step-by-step guardrail creation form in Bolna platform for adding custom safety rules including response triggers, threshold settings, and unwanted utterance detection for Voice AI conversations

Adding your own Guardrail


1

Name

Give your guardrail configuration an identifiable name to easily manage multiple guardrails across your agents.
2

Response

This is the action or message that gets triggered when any unwanted or inappropriate phrases or sentences are detected. Define what your agent should say or do when a guardrail is violated.
3

Threshold

This is the sensitivity level that determines when the system should react to unwanted words or phrases.

If the threshold is set low, the system will react to even slightly inappropriate language. If it’s set higher, only more severe cases will trigger a response. Adjust based on your use case requirements.
4

Utterances

These are the unwanted or inappropriate utterances which you want to guard against. List specific words, phrases, or topics that should trigger your guardrail.

Why use guardrails?

Guardrails help you:
  • Maintain professionalism: Prevent agents from engaging in inappropriate conversations
  • Ensure compliance: Meet industry regulations and content standards
  • Protect brand reputation: Control what your AI agents can discuss
  • Improve user experience: Guide conversations toward productive outcomes
Enhance your agent’s behavior with other Bolna features:
I