Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Jakarta EE 11 Platform launches with modernized Test Compatibility Kit framework

      June 26, 2025

      Can Good UX Protect Older Users From Digital Scams?

      June 25, 2025

      Warp 2.0 evolves terminal experience into an Agentic Development Environment

      June 25, 2025

      Qodo launches CLI agent framework

      June 25, 2025

      My laptop webcam wasn’t cutting it for video calls – then I discovered this accessory

      June 26, 2025

      The top 6 TVs ZDNET readers are buying (no. 1 has the best picture quality we’ve ever seen)

      June 26, 2025

      You should probably delete any sensitive screenshots you have in your phone right now. Here’s why

      June 26, 2025

      Can these $100 Android phones replace my flagship? The result after weeks of testing

      June 26, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      A bot posting the Echo JS RSS feed to Bluesky

      June 26, 2025
      Recent

      A bot posting the Echo JS RSS feed to Bluesky

      June 26, 2025

      Accepting Multiple Parameters in Laravel Commands

      June 26, 2025

      Translate Your App to Other Languages With Laravel Gemini Translator

      June 26, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Distribution Release: deepin 25

      June 26, 2025
      Recent

      Distribution Release: deepin 25

      June 26, 2025

      SpicyPass is a lightweight password manager

      June 26, 2025

      Raspberry Pi 5 Desktop Mini PC: 2.5Gbps Networking

      June 26, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»Tailor responsible AI with new safeguard tiers in Amazon Bedrock Guardrails

    Tailor responsible AI with new safeguard tiers in Amazon Bedrock Guardrails

    June 26, 2025

    Amazon Bedrock Guardrails provides configurable safeguards to help build trusted generative AI applications at scale. It provides organizations with integrated safety and privacy safeguards that work across multiple foundation models (FMs), including models available in Amazon Bedrock, as well as models hosted outside Amazon Bedrock from other model providers and cloud providers. With the standalone ApplyGuardrail API, Amazon Bedrock Guardrails offers a model-agnostic and scalable approach to implementing responsible AI policies for your generative AI applications. Guardrails currently offers six key safeguards: content filters, denied topics, word filters, sensitive information filters, contextual grounding checks, and Automated Reasoning checks (preview), to help prevent unwanted content and align AI interactions with your organization’s responsible AI policies.

    As organizations strive to implement responsible AI practices across diverse use cases, they face the challenge of balancing safety controls with varying performance and language requirements across different applications, making a one-size-fits-all approach ineffective. To address this, we’ve introduced safeguard tiers for Amazon Bedrock Guardrails, so you can choose appropriate safeguards based on your specific needs. For instance, a financial services company can implement comprehensive, multi-language protection for customer-facing AI assistants while using more focused, lower-latency safeguards for internal analytics tools, making sure each application upholds responsible AI principles with the right level of protection without compromising performance or functionality.

    In this post, we introduce the new safeguard tiers available in Amazon Bedrock Guardrails, explain their benefits and use cases, and provide guidance on how to implement and evaluate them in your AI applications.

    Solution overview

    Until now, when using Amazon Bedrock Guardrails, you were provided with a single set of the safeguards associated to specific AWS Regions and a limited set of languages supported. The introduction of safeguard tiers in Amazon Bedrock Guardrails provides three key advantages for implementing AI safety controls:

    • A tier-based approach that gives you control over which guardrail implementations you want to use for content filters and denied topics, so you can select the appropriate protection level for each use case. We provide more details about this in the following sections.
    • Cross-Region Inference Support (CRIS) for Amazon Bedrock Guardrails, so you can use compute capacity across multiple Regions, achieving better scaling and availability for your guardrails. With this, your requests get automatically routed during guardrail policy evaluation to the optimal Region within your geography, maximizing available compute resources and model availability. This helps maintain guardrail performance and reliability when demand increases. There’s no additional cost for using CRIS with Amazon Bedrock Guardrails, and you can select from specific guardrail profiles for controlling model versioning and future upgrades.
    • Advanced capabilities as a configurable tier option for use cases where more robust protection or broader language support are critical priorities, and where you can accommodate a modest latency increase.

    Safeguard tiers are applied at the guardrail policy level, specifically for content filters and denied topics. You can tailor your protection strategy for different aspects of your AI application. Let’s explore the two available tiers:

    • Classic tier (default):
      • Maintains the existing behavior of Amazon Bedrock Guardrails
      • Limited language support: English, French, and Spanish
      • Does not require CRIS for Amazon Bedrock Guardrails
      • Optimized for lower-latency applications
    • Standard tier:
      • Provided as a new capability that you can enable for existing or new guardrails
      • Multilingual support for more than 60 languages
      • Enhanced robustness against prompt typos and manipulated inputs
      • Enhanced prompt attack protection covering modern jailbreak and prompt injection techniques, including token smuggling, AutoDAN, and many-shot, among others
      • Enhanced topic detection with improved understanding and handling of complex topics
      • Requires the use of CRIS for Amazon Bedrock Guardrails and might have a modest increase in latency profile compared to the Classic tier option

    You can select each tier independently for content filters and denied topics policies, allowing for mixed configurations within the same guardrail, as illustrated in the following hierarchy. With this flexibility, companies can implement the right level of protection for each specific application.

    • Policy: Content filters
      • Tier: Classic or Standard
    • Policy: Denied topics
      • Tier: Classic or Standard
    • Other policies: Word filters, sensitive information filters, contextual grounding checks, and Automated Reasoning checks (preview)

    To illustrate how these tiers can be applied, consider a global financial services company deploying AI in both customer-facing and internal applications:

    • For their customer service AI assistant, they might choose the Standard tier for both content filters and denied topics, to provide comprehensive protection across many languages.
    • For internal analytics tools, they could use the Classic tier for content filters prioritizing low latency, while implementing the Standard tier for denied topics to provide robust protection against sensitive financial information disclosure.

    You can configure the safeguard tiers for content filters and denied topics in each guardrail through the AWS Management Console, or programmatically through the Amazon Bedrock SDK and APIs. You can use a new or existing guardrail. For information on how to create or modify a guardrail, see Create your guardrail.

    Your existing guardrails are automatically set to the Classic tier by default to make sure you have no impact on your guardrails’ behavior.

    Quality enhancements with the Standard tier

    According to our tests, the new Standard tier improves harmful content filtering recall by more than 15% with a more than 7% gain in balanced accuracy compared to the Classic tier. A key differentiating feature of the new Standard tier is its multilingual support, maintaining strong performance with over 78% recall and over 88% balanced accuracy for the most common 14 languages.The enhancements in protective capabilities extend across several other aspects. For example, content filters for prompt attacks in the Standard tier show a 30% improvement in recall and 16% gain in balanced accuracy compared to the Classic tier, while maintaining a lower false positive rate. For denied topic detection, the new Standard tier delivers a 32% increase in recall, resulting in an 18% improvement in balanced accuracy.These substantial evolutions in detection capabilities for Amazon Bedrock Guardrails, combined with consistently low false positive rates and robust multilingual performance, also represent a significant advancement in content protection technology compared to other commonly available solutions. The multilingual improvements are particularly noteworthy, with the new Standard tier in Amazon Bedrock Guardrails showing consistent performance gains of 33–49% in recall across different language evaluations compared to other competitors’ options.

    Benefits of safeguard tiers

    Different AI applications have distinct safety requirements based on their audience, content domain, and geographic reach. For example:

    • Customer-facing applications often require stronger protection against potential misuse compared to internal applications
    • Applications serving global customers need guardrails that work effectively across many languages
    • Internal enterprise tools might prioritize controlling specific topics in just a few primary languages

    The combination of the safeguard tiers with CRIS for Amazon Bedrock Guardrails also addresses various operational needs with practical benefits that go beyond feature differences:

    • Independent policy evolution – Each policy (content filters or denied topics) can evolve at its own pace without disrupting the entire guardrail system. You can configure these with specific guardrail profiles in CRIS for controlling model versioning in the models powering your guardrail policies.
    • Controlled adoption – You decide when and how to adopt new capabilities, maintaining stability for production applications. You can continue to use Amazon Bedrock Guardrails with your previous configurations without changes and only move to the new tiers and CRIS configurations when you consider it appropriate.
    • Resource efficiency – You can implement enhanced protections only where needed, balancing security requirements with performance considerations.
    • Simplified migration path – When new capabilities become available, you can evaluate and integrate them gradually by policy area rather than facing all-or-nothing choices. This also simplifies testing and comparison mechanisms such as A/B testing or blue/green deployments for your guardrails.

    This approach helps organizations balance their specific protection requirements with operational considerations in a more nuanced way than a single-option system could provide.

    Configure safeguard tiers on the Amazon Bedrock console

    On the Amazon Bedrock console, you can configure the safeguard tiers for your guardrail in the Content filters tier or Denied topics tier sections by selecting your preferred tier.

    Use of the new Standard tier requires setting up cross-Region inference for Amazon Bedrock Guardrails, choosing the guardrail profile of your choice.

    Configure safeguard tiers using the AWS SDK

    You can also configure the guardrail’s tiers using the AWS SDK. The following is an example to get started with the Python SDK:

    import boto3
    import json
    
    bedrock = boto3.client(
        "bedrock",
        region_name="us-east-1"
    )
    
    # Create a guardrail with Standard tier for both Content Filters and Denied Topics
    response = bedrock.create_guardrail(
        name="enhanced-safety-guardrail",
        # cross-Region is required for STANDARD tier
        crossRegionConfig={
            'guardrailProfileIdentifier': 'us.guardrail.v1:0'
        },
        # Configure Denied Topics with Standard tier
        topicPolicyConfig={
            "topicsConfig": [
                {
                    "name": "Financial Advice",
                    "definition": "Providing specific investment advice or financial recommendations",
                    "type": "DENY",
                    "inputEnabled": True,
                    "inputAction": "BLOCK",
                    "outputEnabled": True,
                    "outputAction": "BLOCK"
                }
            ],
            "tierConfig": {
                "tierName": "STANDARD"
            }
        },
        # Configure Content Filters with Standard tier
        contentPolicyConfig={
            "filtersConfig": [
                {
                    "inputStrength": "HIGH",
                    "outputStrength": "HIGH",
                    "type": "SEXUAL"
                },
                {
                    "inputStrength": "HIGH",
                    "outputStrength": "HIGH",
                    "type": "VIOLENCE"
                }
            ],
            "tierConfig": {
                "tierName": "STANDARD"
            }
        },
        blockedInputMessaging="I cannot respond to that request.",
        blockedOutputsMessaging="I cannot provide that information."
    )

    Within a given guardrail, the content filter and denied topic policies can be configured with its own tier independently, giving you granular control over how guardrails behave. For example, you might choose the Standard tier for content filtering while keeping denied topics in the Classic tier, based on your specific requirements.

    For migrating existing guardrails’ configurations to use the Standard tier, add the sections highlighted in the preceding example for crossRegionConfig and tierConfig to your current guardrail definition. You can do this using the UpdateGuardrail API, or create a new guardrail with the CreateGuardrail API.

    Evaluating your guardrails

    To thoroughly evaluate your guardrails’ performance, consider creating a test dataset that includes the following:

    • Safe examples – Content that should pass through guardrails
    • Harmful examples – Content that should be blocked
    • Edge cases – Content that tests the boundaries of your policies
    • Examples in multiple languages – Especially important when using the Standard tier

    You can also rely on openly available datasets for this purpose. Ideally, your dataset should be labeled with the expected response for each case for assessing accuracy and recall of your guardrails.

    With your dataset ready, you can use the Amazon Bedrock ApplyGuardrail API as shown in the following example to efficiently test your guardrail’s behavior for user inputs without invoking FMs. This way, you can save the costs associated with the large language model (LLM) response generation.

    import boto3
    import json
    
    bedrock_runtime = boto3.client(
        "bedrock-runtime",
        region_name="us-east-1"
    )
    
    # Test the guardrail with potentially problematic content
    content = [
        {
            "text": {
                "text": "Your test prompt here"
            }
        }
    ]
    
    response = bedrock_runtime.apply_guardrail(
        content=content,
        source="INPUT",
        guardrailIdentifier="your-guardrail-id",
        guardrailVersion="DRAFT"
    )
    
    print(json.dumps(response, indent=2, default=str))

    Later, you can repeat the process for the outputs of the LLMs if needed. For this, you can use the ApplyGuardrail API if you want an independent evaluation for models in AWS or outside in another provider, or you can directly use the Converse API if you intend to use models in Amazon Bedrock. When using the Converse API, the inputs and outputs are evaluated with the same invocation request, optimizing latency and reducing coding overheads.

    Because your dataset is labeled, you can directly implement a mechanism for assessing the accuracy, recall, and potential false negatives or false positives through the use of libraries like SKLearn Metrics:

    # scoring script
    # labels and preds store list of ground truth label and guardrails predictions
    
    from sklearn.metrics import confusion_matrix
    
    tn, fp, fn, tp = confusion_matrix(labels, preds, labels=[0, 1]).ravel()
    
    recall = tp / (tp + fn) if (tp + fn) != 0 else 0
    fpr = fp / (fp + tn) if (fp + tn) != 0 else 0
    balanced_accuracy = 0.5 * (recall + 1 - fpr)

    Alternatively, if you don’t have labeled data or your use cases have subjective responses, you can also rely on mechanisms such as LLM-as-a-judge, where you pass the inputs and guardrails’ evaluation outputs to an LLM for assessing a score based on your own predefined criteria. For more information, see Automate building guardrails for Amazon Bedrock using test-drive development.

    Best practices for implementing tiers

    We recommend considering the following aspects when configuring your tiers for Amazon Bedrock Guardrails:

    • Start with staged testing – Test both tiers with a representative sample of your expected inputs and responses before making broad deployment decisions.
    • Consider your language requirements – If your application serves users in multiple languages, the Standard tier’s expanded language support might be essential.
    • Balance safety and performance – Evaluate both the accuracy improvements and latency differences to make informed decisions. Consider if you can afford a few additional milliseconds of latency for improved robustness with the Standard tier or prefer a latency-optimized option for more straight forward evaluations with the Classic tier.
    • Use policy-level tier selection – Take advantage of the ability to select different tiers for different policies to optimize your guardrails. You can choose separate tiers for content filters and denied topics, while combining with the rest of the policies and features available in Amazon Bedrock Guardrails.
    • Remember cross-Region requirements – The Standard tier requires cross-Region inference, so make sure your architecture and compliance requirements can accommodate this. With CRIS, your request originates from the Region where your guardrail is deployed, but it might be served from a different Region from the ones included in the guardrail inference profile for optimizing latency and availability.

    Conclusion

    The introduction of safeguard tiers in Amazon Bedrock Guardrails represents a significant step forward in our commitment to responsible AI. By providing flexible, powerful, and evolving safety tools for generative AI applications, we’re empowering organizations to implement AI solutions that are not only innovative but also ethical and trustworthy. This capabilities-based approach enables you to tailor your responsible AI practices to each specific use case. You can now implement the right level of protection for different applications while creating a path for continuous improvement in AI safety and ethics.The new Standard tier delivers significant improvements in multilingual support and detection accuracy, making it an ideal choice for many applications, especially those serving diverse global audiences or requiring enhanced protection. This aligns with responsible AI principles by making sure AI systems are fair and inclusive across different languages and cultures. Meanwhile, the Classic tier remains available for use cases prioritizing low latency or those with simpler language requirements, allowing organizations to balance performance with protection as needed.

    By offering these customizable protection levels, we’re supporting organizations in their journey to develop and deploy AI responsibly. This approach helps make sure that AI applications are not only powerful and efficient but also align with organizational values, comply with regulations, and maintain user trust.

    To learn more about safeguard tiers in Amazon Bedrock Guardrails, refer to Detect and filter harmful content by using Amazon Bedrock Guardrails, or visit the Amazon Bedrock console to create your first tiered guardrail.


    About the Authors

    Koushik Kethamakka is a Senior Software Engineer at AWS, focusing on AI/ML initiatives. At Amazon, he led real-time ML fraud prevention systems for Amazon.com before moving to AWS to lead development of AI/ML services like Amazon Lex and Amazon Bedrock. His expertise spans product and system design, LLM hosting, evaluations, and fine-tuning. Recently, Koushik’s focus has been on LLM evaluations and safety, leading to the development of products like Amazon Bedrock Evaluations and Amazon Bedrock Guardrails. Prior to joining Amazon, Koushik earned his MS from the University of Houston.

    Hang Su is a Senior Applied Scientist at AWS AI. He has been leading the Amazon Bedrock Guardrails Science team. His interest lies in AI safety topics, including harmful content detection, red-teaming, sensitive information detection, among others.

    Shyam Srinivasan is on the Amazon Bedrock product team. He cares about making the world a better place through technology and loves being part of this journey. In his spare time, Shyam likes to run long distances, travel around the world, and experience new cultures with family and friends.

    Aartika Sardana Chandras is a Senior Product Marketing Manager for AWS Generative AI solutions, with a focus on Amazon Bedrock. She brings over 15 years of experience in product marketing, and is dedicated to empowering customers to navigate the complexities of the AI lifecycle. Aartika is passionate about helping customers leverage powerful AI technologies in an ethical and impactful manner.

    Satveer Khurpa is a Sr. WW Specialist Solutions Architect, Amazon Bedrock at Amazon Web Services, specializing in Amazon Bedrock security. In this role, he uses his expertise in cloud-based architectures to develop innovative generative AI solutions for clients across diverse industries. Satveer’s deep understanding of generative AI technologies and security principles allows him to design scalable, secure, and responsible applications that unlock new business opportunities and drive tangible value while maintaining robust security postures.

    Antonio Rodriguez is a Principal Generative AI Specialist Solutions Architect at Amazon Web Services. He helps companies of all sizes solve their challenges, embrace innovation, and create new business opportunities with Amazon Bedrock. Apart from work, he loves to spend time with his family and play sports with his friends.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleFrom Interaction to Impact: Towards Safer AI Agents Through Understanding and Evaluating Mobile UI Operation Impacts
    Next Article Structured data response with Amazon Bedrock: Prompt Engineering and Tool Use

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    June 26, 2025
    Machine Learning

    Using Amazon SageMaker AI Random Cut Forest for NASA’s Blue Origin spacecraft sensor data

    June 26, 2025
    Leave A Reply Cancel Reply

    For security, use of Google's reCAPTCHA service is required which is subject to the Google Privacy Policy and Terms of Use.

    Continue Reading

    CQRLOG – advanced ham radio logger

    Linux

    CVE-2025-3974 – PHPGurukul COVID19 Testing Management System SQL Injection Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    Development Release: Emmabuntüs DE6 RC1

    News & Updates

    Google DeepMind at NeurIPS 2024

    Artificial Intelligence

    Highlights

    CVE-2025-5693 – PHPGurukul Human Metapneumovirus Testing Management System SQL Injection

    June 5, 2025

    CVE ID : CVE-2025-5693

    Published : June 5, 2025, 8:15 p.m. | 1 hour, 52 minutes ago

    Description : A vulnerability was found in PHPGurukul Human Metapneumovirus Testing Management System 1.0. It has been declared as critical. Affected by this vulnerability is an unknown functionality of the file /bwdates-report-result.php. The manipulation of the argument fromdate/todate leads to sql injection. The attack can be launched remotely. The exploit has been disclosed to the public and may be used.

    Severity: 6.3 | MEDIUM

    Visit the link for more details, such as CVSS details, affected products, timeline, and more…

    What is Artificial Intelligence? Everything You Need to Know

    May 6, 2025

    Why do Xbox’s custom controllers go so unreasonably hard?!

    June 3, 2025

    Cisco ISE Vulnerability Allows Remote to Access Sensitive Data – PoC Exploit Available

    June 5, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.