Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      15 Proven Benefits of Outsourcing Node.js Development for Large Organizations

      July 9, 2025

      10 Reasons to Choose Full-Stack Techies for Your Next React.js Development Project

      July 9, 2025

      Anthropic proposes transparency framework for frontier AI development

      July 8, 2025

      Sonatype Open Source Malware Index, Gemini API Batch Mode, and more – Daily News Digest

      July 8, 2025

      Microsoft sees its carbon emissions soar on a 168% glut in AI energy demand, “we recognize that we must also bring more carbon-free electricity onto the grids.”

      July 9, 2025

      You can get a Snapdragon X-powered laptop for under $500 right now — a low I didn’t think we’d see this Prime Day week

      July 9, 2025

      Sam Altman admits current computers were designed for an AI-free world — but OpenAI’s new type of computer will make the AI revolution “transcendentally good”

      July 9, 2025

      It doesn’t matter how many laptops I review or how great the deals are — this is the one I keep coming back to over and over again

      July 9, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Leading Experts in Meme Coin Development – Beleaf Technologies

      July 9, 2025
      Recent

      Leading Experts in Meme Coin Development – Beleaf Technologies

      July 9, 2025

      Redefining Quality Engineering – Tricentis India Partner Event

      July 9, 2025

      Enhancing JSON Responses with Laravel Model Appends

      July 9, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Microsoft sees its carbon emissions soar on a 168% glut in AI energy demand, “we recognize that we must also bring more carbon-free electricity onto the grids.”

      July 9, 2025
      Recent

      Microsoft sees its carbon emissions soar on a 168% glut in AI energy demand, “we recognize that we must also bring more carbon-free electricity onto the grids.”

      July 9, 2025

      You can get a Snapdragon X-powered laptop for under $500 right now — a low I didn’t think we’d see this Prime Day week

      July 9, 2025

      Sam Altman admits current computers were designed for an AI-free world — but OpenAI’s new type of computer will make the AI revolution “transcendentally good”

      July 9, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»Configure fine-grained access to Amazon Bedrock models using Amazon SageMaker Unified Studio

    Configure fine-grained access to Amazon Bedrock models using Amazon SageMaker Unified Studio

    July 9, 2025

    Enterprises adopting advanced AI solutions recognize that robust security and precise access control are essential for protecting valuable data, maintaining compliance, and preserving user trust. As organizations expand AI usage across teams and applications, they require granular permissions to safeguard sensitive information and manage who can access powerful models. Amazon SageMaker Unified Studio addresses these needs so organizations can configure fine-grained access policies, making sure that only authorized users can interact with foundation models (FMs) while supporting secure, collaborative innovation at scale.

    Launched in 2025, SageMaker Unified Studio is a single data and AI development environment where you can find and access the data in your organization and act on it using the best tools across use cases. SageMaker Unified Studio brings together the functionality and tools from existing AWS analytics and AI/ML services, including Amazon EMR, AWS Glue, Amazon Athena, Amazon Redshift, Amazon Bedrock, and Amazon SageMaker AI.

    Amazon Bedrock in SageMaker Unified Studio provides various options for discovering and experimenting with Amazon Bedrock models and applications. For example, you can use a chat playground to try a prompt with Anthropic’s Claude without having to write code. You can also create a generative AI application that uses an Amazon Bedrock model and features, such as a knowledge base or a guardrail. To learn more, refer to Amazon Bedrock in SageMaker Unified Studio.

    In this post, we demonstrate how to use SageMaker Unified Studio and AWS Identity and Access Management (IAM) to establish a robust permission framework for Amazon Bedrock models. We show how administrators can precisely manage which users and teams have access to specific models within a secure, collaborative environment. We guide you through creating granular permissions to control model access, with code examples for common enterprise governance scenarios. By the end, you will understand how to tailor access to generative AI capabilities to meet your organization’s requirements—addressing a core challenge in enterprise AI adoption by enabling developer flexibility while maintaining strong security standards.

    Solution overview

    In SageMaker Unified Studio, a domain serves as the primary organizational structure, so you can oversee multiple AWS Regions, accounts, and workloads from a single interface. Each domain is assigned a unique URL and offers centralized control over studio configurations, user accounts, and network settings.

    Inside each domain, projects facilitate streamlined collaboration. Projects can span different Regions or accounts within a Region, and their metadata includes details such as the associated Git repository, team members, and their permissions. Every account where a project has resources is assigned at least one project role, which determines the tools, compute resources, datasets, and artificial intelligence and machine learning (AI/ML) assets accessible to project members. To manage data access, you can adjust the IAM permissions tied to the project’s role. SageMaker Unified Studio uses several IAM roles. For a complete list, refer to Identity and access management for Amazon SageMaker Unified Studio.

    There are two primary methods for users to interact with Amazon Bedrock models in SageMaker Unified Studio: the SageMaker Unified Studio playground and SageMaker Unified Studio projects.

    In the SageMaker Unified Studio playground scenario, model consumption roles provide secure access to Amazon Bedrock FMs. You can choose between automatic role creation for individual models or configuring a single role for all models. The default AmazonSageMakerBedrockModelConsumptionRole comes with preconfigured permissions to consume Amazon Bedrock models, including invoking the Amazon Bedrock application inference profile created for a particular SageMaker Unified Studio domain. To fine-tune access control, you can add inline policies to these roles that explicitly allow or deny access to specific Amazon Bedrock models.

    The following diagram illustrates this architecture.

    Playground Access

    The workflow consists of the following steps:

    1. Initial access path:
      1. The SageMakerDomain execution role connects to the SageMaker Unified Studio domain.
      2. The connection flows to the SageMaker user profile.
      3. The user profile accesses SageMaker Unified Studio.
    2. Restricted access path (top flow):
      1. The direct access attempt from SageMaker Unified Studio to Amazon Bedrock is denied.
      2. The IAM policy blocks access to FMs.
      3. Anthropic’s Claude 3.7 Sonnet access is denied (marked with X).
      4. Anthropic’s Claude 3.5 Haiku access is denied (marked with X).
    3. Permitted access path (bottom flow):
      1. SageMakerBedrockModelConsumptionRole is used.
      2. The appropriate IAM policy allows access (marked with a checkmark).
      3. Amazon Bedrock access is permitted.
      4. Anthropic’s Claude 3.7 Sonnet access is allowed (marked with checkmark).
      5. Anthropic’s Claude 3.5 Haiku access is allowed (marked with checkmark).
    4. Governance mechanism:
      1. IAM policies serve as the control point for model access.
      2. Different roles determine different levels of access permission.
      3. Access controls are explicitly defined for each FM.

    In the SageMaker Unified Studio project scenario, SageMaker Unified Studio uses a model provisioning role to create an inference profile for an Amazon Bedrock model in a project. The inference profile is required for the project to interact with the model. You can either let SageMaker Unified Studio automatically create a unique model provisioning role, or you can provide a custom model provisioning role. The default AmazonSageMakerBedrockModelManagementRole has the AWS policy AmazonDataZoneBedrockModelManagementPolicy attached. You can restrict access to specific account IDs through custom trust policies. You can also attach inline policies and use the statement CreateApplicationInferenceProfileUsingFoundationModels to allow or deny access to specific Amazon Bedrock models in your project.

    The following diagram illustrates this architecture.

    Projects Access

    The workflow consists of the following steps:

    1. Initial access path:
      1. The SageMakerDomain execution role connects to the SageMaker Unified Studio domain.
      2. The connection flows to the SageMaker user profile.
      3. The user profile accesses SageMaker Unified Studio.
    2. Restricted access path (top flow):
      1. The direct access attempt from SageMaker Unified Studio to Amazon Bedrock is denied.
      2. The IAM policy blocks access to FMs.
      3. Anthropic’s Claude 3.7 Sonnet access is denied (marked with X).
      4. Anthropic’s Claude 3.5 Haiku access is denied (marked with X).
    3. Permitted access path (bottom flow):
      1. SageMakerBedrockModelManagementRole is used.
      2. The appropriate IAM policy allows access (marked with a checkmark).
      3. Amazon Bedrock access is permitted.
      4. Anthropic’s Claude 3.7 Sonnet access is allowed (marked with checkmark).
      5. Anthropic’s Claude 3.5 Haiku access is allowed (marked with checkmark).
    4. Governance mechanism:
      1. IAM policies serve as the control point for model access.
      2. Different roles determine different levels of access permission.
      3. Access controls are explicitly defined for each FM.

    By customizing the policies attached to these roles, you can control which actions are permitted or denied, thereby governing access to generative AI capabilities.

    To use a specific model from Amazon Bedrock, SageMaker Unified Studio uses the model ID of the chosen model as part of the API calls. At the time of writing, SageMaker Unified Studio supports the following Amazon Bedrock models (the full list of current models can be found here), grouped by model provider:

    • Amazon:
      • Amazon Titan Text G1 – Premier: amazon.titan-text-premier-v1:0
      • Amazon Nova Pro: amazon.nova-pro-v1:0
      • Amazon Nova Lite: amazon.nova-lite-v1:0
      • Amazon Nova Canvas: amazon.nova-canvas-v1:0
    • Stability AI:
      • SDXL 1.0: stability.stable-diffusion-xl-v1
    • AI21 Labs:
      • Jamba-Instruct: ai21.jamba-instruct-v1:0
      • Jamba 1.5 Large: ai21.jamba-1-5-large-v1:0
      • Jamba 1.5 Mini: ai21.jamba-1-5-mini-v1:0
    • Anthropic:
      • Claude 3.7 Sonnet: anthropic.claude-3-7-sonnet-20250219-v1:0
    • Cohere:
      • Command R+: cohere.command-r-plus-v1:0
      • Command Light: cohere.command-light-text-v14
      • Embed Multilingual: cohere.embed-multilingual-v3
    • DeepSeek:
      • DeepSeek-R1: deepseek.r1-v1:0
    • Meta:
      • Llama 3.3 70B Instruct: meta.llama3-3-70b-instruct-v1:0
      • Llama 4 Scout 17B Instruct: meta.llama4-scout-17b-instruct-v1:0
      • Llama 4 Maverick 17B Instruct: meta.llama4-maverick-17b-instruct-v1:0
    • Mistral AI:
      • Mistral 7B Instruct: mistral.mistral-7b-instruct-v0:2
      • Pixtral Large (25.02): mistral.pixtral-large-2502-v1:0

    Create a model consumption role for the playground scenario

    In the following steps, you create an IAM role with a trust policy, add two inline policies, and attach them to the role.

    Create the IAM role with a trust policy

    Complete the following steps to create an IAM role with a trust policy:

    1. On the IAM console, in the navigation pane, choose Roles, then choose Create role.
    2. For Trusted entity type, select Custom trust policy.
    3. Delete the default policy in the editor and enter the following trust policy (replace account-id for the aws:SourceAccount field with your AWS account ID):
    {
        "Version": "2012-10-17",
        "Statement": [
            {
                "Effect": "Allow",
                "Principal": {
                    "Service": "datazone.amazonaws.com"
                },
                "Action": [
                    "sts:AssumeRole",
                    "sts:SetContext"
                ],
                "Condition": {
                    "StringEquals": {
                        "aws:SourceAccount": "[account-id]"
                    }
                }
            }
        ]
    }
    1. Choose Next.
    2. Skip the Add permissions page by choosing Next.
    3. Enter a name for the role (for example, DataZoneBedrock-Role) and an optional description.
    4. Choose Create role.

    Add the first inline policy

    Complete the following steps to add an inline policy:

    1. On the IAM console, open the newly created role details page.
    2. On the Permissions tab, choose Add permissions and then Create inline policy.
    3. On the JSON tab, delete the default policy and enter the first inline policy:
    {
        "Version": "2012-10-17",
        "Statement": [
            {
                "Sid": "InvokeDomainInferenceProfiles",
                "Effect": "Allow",
                "Action": [
                    "bedrock:InvokeModel",
                    "bedrock:InvokeModelWithResponseStream"
                ],
                "Resource": "arn:aws:bedrock:*:*:application-inference-profile/*",
                "Condition": {
                    "StringEquals": {
                        "aws:ResourceTag/AmazonDataZoneDomain": "${datazone:domainId}",
                        "aws:ResourceAccount": "${aws:PrincipalAccount}"
                    },
                    "Null": {
                        "aws:ResourceTag/AmazonDataZoneProject": "true"
                    }
                }
            }
        ]
    }
    1. Choose Review policy.
    2. Name the policy (for example, DataZoneDomainInferencePolicy) and choose Create policy.

    Add the second inline policy

    Complete the following steps to add another inline policy:

    1. On the role’s Permissions tab, choose Add permissions and then Create inline policy.
    2. On the JSON tab, delete the default policy and enter the second inline policy (replace account-id in the bedrock:InferenceProfileArn field with your account ID):
    {
        "Version": "2012-10-17",
        "Statement": [
            {
                "Sid": "AllowInferenceProfileToInvokeFoundationModels",
                "Effect": "Allow",
                "Action": [
                    "bedrock:InvokeModel",
                    "bedrock:InvokeModelWithResponseStream"
                ],
                "Resource": [
                    "arn:aws:bedrock:us-east-1::foundation-model/anthropic.claude-3-5-haiku-20241022-v1:0",
                    "arn:aws:bedrock:us-east-2::foundation-model/anthropic.claude-3-5-haiku-20241022-v1:0",
                    "arn:aws:bedrock:us-west-2::foundation-model/anthropic.claude-3-5-haiku-20241022-v1:0"
                ],
                "Condition": {
                    "ArnLike": {
                        "bedrock:InferenceProfileArn": "arn:aws:bedrock:*:[account-id]:application-inference-profile/*"
                    }
                }
            }
        ]
    }
    1. Choose Review policy.
    2. Name the policy (for example, BedrockFoundationModelAccessPolicy) and choose Create policy.

    Explanation of the policies

    In this section, we discuss the details of the policies.

    Trust policy explanation

    This trust policy defines who can assume the IAM role:

    • It allows the Amazon DataZone service (datazone.amazonaws.com) to assume this role
    • The service can perform sts:AssumeRole and sts:SetContext actions
    • A condition restricts this to only when the AWS account you specify is your AWS account

    This makes sure that only Amazon DataZone from your specified account can assume this role.

    First inline policy explanation

    This policy controls access to Amazon Bedrock inference profiles:

    • It allows invoking Amazon Bedrock models (bedrock:InvokeModel and bedrock:InvokeModelWithResponseStream), but only for resources that are application inference profiles (arn:aws:bedrock:::application-inference-profile/*)
    • It has three important conditions:
    • The profile must be tagged with AmazonDataZoneDomain matching the domain ID of the caller
    • The resource must be in the same AWS account as the principal making the request
    • The resource must not have an AmazonDataZoneProject tag set to "true"

    This effectively limits access to only those inference profiles that belong to the same Amazon DataZone domain as the caller and are not associated with a specific project.

    Second inline policy explanation

    This policy controls which specific FMs can be accessed:

    • It allows the same Amazon Bedrock model invocation actions, but only for specific Anthropic Claude 3.5 Haiku models in three Regions:
      • us-east-1
      • us-east-2
      • us-west-2
    • It has a condition that the request must come through an inference profile from your account

    Combined effect on the SageMaker Unified Studio domain generative AI playground

    Together, these policies create a secure, controlled environment for using Amazon Bedrock models in the SageMaker Unified Studio domain through the following methods:

    • Limiting model access – Only the specified Anthropic Claude 3.5 Haiku model can be used, not other Amazon Bedrock models
    • Enforcing access through inference profiles – Models can only be accessed through properly configured application inference profiles
    • Maintaining domain isolation – Access is restricted to inference profiles tagged with the user’s Amazon DataZone domain
    • Helping to prevent cross-account access – Resources must be in the same account as the principal
    • Regional restrictions – Access the model is only allowed in three specific AWS Regions

    This implementation follows the principle of least privilege by providing only the minimum permissions needed for the intended use case, while maintaining proper security boundaries between different domains and projects.

    Create a model provisioning role for the project scenario

    In this section, we walk through the steps to create an IAM role with a trust policy and add the required inline policy to make sure that the models are limited to the approved ones.

    Create the IAM role with a trust policy

    Complete the following steps to create an IAM role with a trust policy:

    1. On the IAM console, in the navigation pane, choose Roles, then choose Create role.
    2. For Trusted entity type, select Custom trust policy.
    3. Delete the default policy in the editor and enter the following trust policy (replace account-id for the aws:SourceAccount field with your account ID):
    {
        "Version": "2012-10-17",
        "Statement": [
            {
                "Effect": "Allow",
                "Principal": {
                    "Service": "datazone.amazonaws.com"
                },
                "Action": [
                    "sts:AssumeRole",
                    "sts:SetContext"
                ],
                "Condition": {
                    "StringEquals": {
                        "aws:SourceAccount": "[account-id]"
                    }
                }
            }
        ]
    }
    1. Choose Next.
    2. Skip the Add permissions page by choosing Next.
    3. Enter a name for the role (for example, SageMakerModelManagementRole) and an optional description, such as Role for managing Bedrock model access in SageMaker Unified Studio.
    4. Choose Create role.

    Add the inline policy

    Complete the following steps to add an inline policy:

    1. On the IAM console, open the details page of the newly created role.
    2. On the Permissions tab, choose Add permissions and then Create inline policy.
    3. On the JSON tab, delete the default policy and enter the following inline policy:
    {
        "Version": "2012-10-17",
        "Statement": [
            {
                "Sid": "ManageApplicationInferenceProfile",
                "Effect": "Allow",
                "Action": [
                    "bedrock:CreateInferenceProfile",
                    "bedrock:TagResource"
                ],
                "Resource": [
                    "arn:aws:bedrock:*:*:application-inference-profile/*"
                ],
                "Condition": {
                    "StringEquals": {
                        "aws:ResourceAccount": "${aws:PrincipalAccount}"
                    },
                    "ForAnyValue:StringEquals": {
                        "aws:TagKeys": [
                            "AmazonDataZoneProject"
                        ]
                    },
                    "Null": {
                        "aws:ResourceTag/AmazonDataZoneProject": "false",
                        "aws:RequestTag/AmazonDataZoneProject": "false"
                    }
                }
            },
            {
                "Sid": "DeleteApplicationInferenceProfile",
                "Effect": "Allow",
                "Action": [
                    "bedrock:DeleteInferenceProfile"
                ],
                "Resource": [
                    "arn:aws:bedrock:*:*:application-inference-profile/*"
                ],
                "Condition": {
                    "StringEquals": {
                        "aws:ResourceAccount": "${aws:PrincipalAccount}"
                    },
                    "Null": {
                        "aws:ResourceTag/AmazonDataZoneProject": "false"
                    }
                }
            },
            {
                "Sid": "CreateApplicationInferenceProfileUsingFoundationModels",
                "Effect": "Allow",
                "Action": [
                    "bedrock:CreateInferenceProfile"
                ],
                "Resource": [
                    "arn:aws:bedrock:*::foundation-model/anthropic.claude-3-5-haiku-20241022-v1:0"
                ]
            },
            {
                "Sid": "CreateApplicationInferenceProfileUsingBedrockModels",
                "Effect": "Allow",
                "Action": [
                    "bedrock:CreateInferenceProfile"
                ],
                "Resource": [
                    "arn:aws:bedrock:*:*:inference-profile/*"
                ],
                "Condition": {
                    "StringEquals": {
                        "aws:ResourceAccount": "${aws:PrincipalAccount}"
                    }
                }
            }
        ]
    }
    1. Choose Review policy.
    2. Name the policy (for example, BedrockModelManagementPolicy) and choose Create policy.

    Explanation of the policies

    In this section, we discuss the details of the policies.

    Trust policy explanation

    This trust policy defines who can assume the IAM role:

    • It allows the Amazon DataZone service to assume this role
    • The service can perform sts:AssumeRole and sts:SetContext actions
    • A condition restricts this to only when the source AWS account is your AWS account

    This makes sure that only Amazon DataZone from your specific account can assume this role.

    Inline policy explanation

    This policy controls access to Amazon Bedrock inference profiles and models:

    • It allows creating and managing application inference profiles (bedrock:CreateInferenceProfile, bedrock:TagResource, bedrock:DeleteInferenceProfile)
    • It specifically permits creating inference profiles for only the anthropic.claude-3-5-haiku-20241022-v1:0 model
    • Access is controlled through several conditions:
      • Resources must be in the same AWS account as the principal making the request
      • Operations are restricted based on the AmazonDataZoneProject tag
      • Creating profiles requires proper tagging with AmazonDataZoneProject
      • Deleting profiles is only allowed for resources with the appropriate AmazonDataZoneProject tag

    Combined effect on SageMaker Unified Studio domain project

    Together, these policies create a secure, controlled environment for using Amazon Bedrock models in SageMaker Unified Studio domain projects through the following methods:

    • Limiting model access – Only the specified Anthropic Claude 3.5 Haiku model can be used
    • Enforcing proper tagging – Resources must be properly tagged with AmazonDataZoneProject identifiers
    • Maintaining account isolation – Resources must be in the same account as the principal
    • Implementing least privilege – Only specific actions (create, tag, delete) are permitted on inference profiles
    • Providing project-level isolation – Access is restricted to inference profiles tagged with appropriate project identifiers

    This implementation follows the principle of least privilege by providing only the minimum permissions needed for project-specific use cases, while maintaining proper security boundaries between different projects and making sure that only approved FMs can be accessed.

    Configure a SageMaker Unified Studio domain to use the roles

    Complete the following steps to create a SageMaker Unified Studio domain and configure it to use the roles you created:

    1. On the SageMaker console, choose the appropriate Region.
    2. Choose Create a Unified Studio domain and then choose Quick setup.
    3. For Name, enter a meaningful domain name.
    4. Scroll down to Generative AI resources.
    5. Under Model provisioning role, choose the model management role created in the previous section.
    6. Under Model consumption role, select Use a single existing role for all models and choose the model consumption role created in the previous section.
    7. Complete the remaining steps according to your AWS IAM Identity Center configurations and create your domain.

    Clean up

    To avoid incurring future charges related to various services used in SageMaker Unified Studio, log out of the SageMaker Unified Studio domain and delete the domain in SageMaker Unified Studio.

    Conclusion

    In this post, we demonstrated how the SageMaker Unified Studio playground and SageMaker Unified Studio projects invoke large language models powered by Amazon Bedrock, and how enterprises can govern access to these models, whether you want to limit access to specific models or to every model from the service. You can combine the IAM policies shown in this post in the same IAM role to provide complete control. By following these guidelines, enterprises can make sure their use of generative AI models is both secure and aligned with organizational policies. This approach not only safeguards sensitive data but also empowers business analysts and data scientists to harness the full potential of AI within a controlled environment.

    Now that your environment is configured with strong identity based policies, we suggest reading the following posts to learn how Amazon SageMaker Unified Studio enables you to securely innovate quickly, and at scale, with generative AI:

    • An integrated experience for all your data and AI with Amazon SageMaker Unified Studio
    • Unified scheduling for visual ETL flows and query books in Amazon SageMaker Unified Studio
    • Foundational blocks of Amazon SageMaker Unified Studio: An admin’s guide to implement unified access to all your data, analytics, and AI

    About the authors

    VarunVarun Jasti is a Solutions Architect at Amazon Web Services, working with AWS Partners to design and scale artificial intelligence solutions for public sector use cases to meet compliance standards. With a background in Computer Science, his work covers broad range of ML use cases primarily focusing on LLM training/inferencing and computer vision. In his spare time, he loves playing tennis and swimming.

    Saptarshi BanarjeeSaptarshi Banarjee serves as a Senior Solutions Architect at AWS, collaborating closely with AWS Partners to design and architect mission-critical solutions. With a specialization in generative AI, AI/ML, serverless architecture, Next-Gen Developer Experience tools and cloud-based solutions, Saptarshi is dedicated to enhancing performance, innovation, scalability, and cost-efficiency for AWS Partners within the cloud ecosystem.

    JonJon Turdiev is a Senior Solutions Architect at Amazon Web Services, where he helps startup customers build well-architected products in the cloud. With over 20 years of experience creating innovative solutions in cybersecurity, AI/ML, healthcare, and Internet of Things (IoT), Jon brings deep technical expertise to his role. Previously, Jon founded Zehntec, a technology consulting company, and developed award-winning medical bedside terminals deployed in hospitals worldwide. Jon holds a Master’s degree in Computer Science and shares his knowledge through webinars, workshops, and as a judge at hackathons.

    LijanLijan Kuniyil is a Senior Technical Account Manager at AWS. Lijan enjoys helping AWS enterprise customers build highly reliable and cost-effective systems with operational excellence. Lijan has over 25 years of experience in developing solutions for financial, healthcare and consulting companies.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleQuery Amazon Aurora PostgreSQL using Amazon Bedrock Knowledge Bases structured data
    Next Article Improve conversational AI response times for enterprise applications with the Amazon Bedrock streaming API and AWS AppSync

    Related Posts

    Machine Learning

    Improve conversational AI response times for enterprise applications with the Amazon Bedrock streaming API and AWS AppSync

    July 9, 2025
    Machine Learning

    Query Amazon Aurora PostgreSQL using Amazon Bedrock Knowledge Bases structured data

    July 9, 2025
    Leave A Reply Cancel Reply

    For security, use of Google's reCAPTCHA service is required which is subject to the Google Privacy Policy and Terms of Use.

    Continue Reading

    CVE-2025-20325 – Splunk Enterprise and Cloud Platform Search Head Cluster Secret Key Exposure

    Common Vulnerabilities and Exposures (CVEs)

    CVE-2025-38172 – “Linux EROFS UAF Vulnerability”

    Common Vulnerabilities and Exposures (CVEs)

    CodeSOD: A Ruby Encrusted Footgun

    News & Updates

    SmartOS – Type 1 Hypervisor platform based on illumos

    Linux

    Highlights

    10 Best Free and Open Source Collection Managers

    April 30, 2025

    There is a wide range of collection manager software available for Linux. We recommend the…

    CVE-2025-47658 – ELEX WordPress HelpDesk & Customer Ticketing System Unrestricted File Upload RCE

    May 27, 2025

    Duolingo just added 148 new courses in its biggest update ever – thanks to AI

    April 30, 2025

    How to Turn Ubuntu 24.04 into a KVM Hypervisor – Quick Setup with Web Management

    April 28, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.