AWS Guardrails

AWS Guardrails provides configurable security measures to help safely build generative AI applications at scale. It uses automated reasoning to help prevent factual errors caused by hallucinations. It

AWS Guardrails features include:

  • Content filtering (preventing explicit, violent content, etc.)

  • Topic deviation prevention

  • Word restrictions

  • Sensitive data removal

  • Hallucination prevention

Operation Steps

Step 1: Configure Settings (Name, Description, Blocked Message Response)

Example settings:

Step 2: Set Content Filtering (User Input, Assistant Response)

You can specify harm levels, including image inputs

Example settings:

Step 3: Topic Rejection, Preventing Off-Topic Responses

You can prevent off-topic responses by adding descriptions, definitions, and examples

Example settings:

Step 4: Word Restrictions (Blocking Profanity or Specific Terms)

For example: To prevent the model from mentioning Voldemort, add words - multiple words can be input at once (csv, txt, ctrl + c).

Example settings:

Step 5: Sensitive Data (PII) Blocking and Regex Replacement

Several patterns are preset, such as credit cards, vehicle numbers, ID numbers, etc.

Example settings:

Regex replacement allows you to replace unwanted structures in AI assistant responses using desired regular expressions.

Step 6: Preventing Hallucinations and Setting Relevance Threshold for Response Blocking

Example settings:

Last updated

Was this helpful?