Close Menu
    DevStackTipsDevStackTips
    • Home
    • News & Updates
      1. Tech & Work
      2. View All

      Designing For TV: The Evergreen Pattern That Shapes TV Experiences

      August 27, 2025

      Amplitude launches new self-service capabilities for marketing initiatives

      August 27, 2025

      Microsoft packs Visual Studio August update with smarter AI features

      August 27, 2025

      Optimizing PWAs For Different Display Modes

      August 26, 2025

      Why this $25 ratchet tool beats any multitool or Swiss Army Knife I’ve ever tested

      August 28, 2025

      One of my favorite sports watches from 2024 just got upgrades in all the right places

      August 28, 2025

      Google’s AI Mode is getting more links for you not to click on

      August 28, 2025

      I still prefer Apple Watch over Oura Ring for 3 key reasons – but there is one big drawback

      August 28, 2025
    • Development
      1. Algorithms & Data Structures
      2. Artificial Intelligence
      3. Back-End Development
      4. Databases
      5. Front-End Development
      6. Libraries & Frameworks
      7. Machine Learning
      8. Security
      9. Software Engineering
      10. Tools & IDEs
      11. Web Design
      12. Web Development
      13. Web Security
      14. Programming Languages
        • PHP
        • JavaScript
      Featured

      Heartbeat Collection Method in Laravel 12.26; Wayfinder Now in React and Vue Starter Kits

      August 28, 2025
      Recent

      Heartbeat Collection Method in Laravel 12.26; Wayfinder Now in React and Vue Starter Kits

      August 28, 2025

      spatie/laravel-rdap

      August 28, 2025

      mvanduijker/laravel-mercure-broadcaster

      August 28, 2025
    • Operating Systems
      1. Windows
      2. Linux
      3. macOS
      Featured

      Geekom’s A9 Max mini PC is so good that I want to turn off my desktop gaming rig — and it’s not bad at AI, either

      August 28, 2025
      Recent

      Geekom’s A9 Max mini PC is so good that I want to turn off my desktop gaming rig — and it’s not bad at AI, either

      August 28, 2025

      ‘There Are No Ghosts At The Grand’ looks glorious — I’m more excited than ever for this upcoming Xbox Game Pass release

      August 28, 2025

      Epic Games CEO Tim Sweeney says Unreal Engine 5’s performance problems aren’t about the engine — they’re about when developers choose to optimize

      August 28, 2025
    • Learning Resources
      • Books
      • Cheatsheets
      • Tutorials & Guides
    Home»Development»Machine Learning»Mercury foundation models from Inception Labs are now available in Amazon Bedrock Marketplace and Amazon SageMaker JumpStart

    Mercury foundation models from Inception Labs are now available in Amazon Bedrock Marketplace and Amazon SageMaker JumpStart

    August 28, 2025

    Today, we are excited to announce that Mercury and Mercury Coder foundation models (FMs) from Inception Labs are available through Amazon Bedrock Marketplace and Amazon SageMaker JumpStart. With this launch, you can deploy the Mercury FMs to build, experiment, and responsibly scale your generative AI applications on AWS.

    In this post, we demonstrate how to get started with Mercury models on Amazon Bedrock Marketplace and SageMaker JumpStart.

    About Mercury foundation models

    Mercury is the first family of commercial-scale diffusion-based language models, offering groundbreaking advancements in generation speed while maintaining high-quality outputs. Unlike traditional autoregressive models that generate text one token at a time, Mercury models use diffusion to generate multiple tokens in parallel through a coarse-to-fine approach, resulting in dramatically faster inference speeds. Mercury Coder models deliver the following key features:

    • Ultra-fast generation speeds of up to 1,100 tokens per second on NVIDIA H100 GPUs, up to 10 times faster than comparable models
    • High-quality code generation across multiple programming languages, including Python, Java, JavaScript, C++, PHP, Bash, and TypeScript
    • Strong performance on fill-in-the-middle tasks, making them ideal for code completion and editing workflows
    • Transformer-based architecture, providing compatibility with existing optimization techniques and infrastructure
    • Context length support of up to 32,768 tokens out of the box and up to 128,000 tokens with context extension approaches

    About Amazon Bedrock Marketplace

    Amazon Bedrock Marketplace plays a pivotal role in democratizing access to advanced AI capabilities through several key advantages:

    • Comprehensive model selection – Amazon Bedrock Marketplace offers an exceptional range of models, from proprietary to publicly available options, so organizations can find the perfect fit for their specific use cases.
    • Unified and secure experience – By providing a single access point for models through the Amazon Bedrock APIs, Amazon Bedrock Marketplace significantly simplifies the integration process. Organizations can use these models securely, and for models that are compatible with the Amazon Bedrock Converse API, you can use the robust toolkit of Amazon Bedrock, including Amazon Bedrock Agents, Amazon Bedrock Knowledge Bases, Amazon Bedrock Guardrails, and Amazon Bedrock Flows.
    • Scalable infrastructure – Amazon Bedrock Marketplace offers configurable scalability through managed endpoints, so organizations can select their desired number of instances, choose appropriate instance types, define custom automatic scaling policies that dynamically adjust to workload demands, and optimize costs while maintaining performance.

    Deploy Mercury and Mercury Coder models in Amazon Bedrock Marketplace

    Amazon Bedrock Marketplace gives you access to over 100 popular, emerging, and specialized foundation models through Amazon Bedrock. To access the Mercury models in Amazon Bedrock, complete the following steps:

    1. On the Amazon Bedrock console, in the navigation pane under Foundation models, choose Model catalog.

    You can also use the Converse API to invoke the model with Amazon Bedrock tooling.

    1. On the Model catalog page, filter for Inception as a provider and choose the Mercury model.

    The Model detail page provides essential information about the model’s capabilities, pricing structure, and implementation guidelines. You can find detailed usage instructions, including sample API calls and code snippets for integration.

    1. To begin using the Mercury model, choose Subscribe.

    1. On the model detail page, choose Deploy.

    You will be prompted to configure the deployment details for the model. The model ID will be prepopulated.

    1. For Endpoint name, enter an endpoint name (between 1–50 alphanumeric characters).
    2. For Number of instances, enter a number of instances (between 1–100).
    3. For Instance type, choose your instance type. For optimal performance with Nemotron Super, a GPU-based instance type like ml.p5.48xlarge is recommended.
    4. Optionally, you can configure advanced security and infrastructure settings, including virtual private cloud (VPC) networking, service role permissions, and encryption settings. For most use cases, the default settings will work well. However, for production deployments, you might want to review these settings to align with your organization’s security and compliance requirements.
    5. Choose Deploy to begin using the model.

    When the deployment is complete, you can test its capabilities directly in the Amazon Bedrock playground.This is an excellent way to explore the model’s reasoning and text generation abilities before integrating it into your applications. The playground provides immediate feedback, helping you understand how the model responds to various inputs and letting you fine-tune your prompts for optimal results. You can use these models with the Amazon Bedrock Converse API.

    SageMaker JumpStart overview

    SageMaker JumpStart is a fully managed service that offers state-of-the-art FMs for various use cases such as content writing, code generation, question answering, copywriting, summarization, classification, and information retrieval. It provides a collection of pre-trained models that you can deploy quickly, accelerating the development and deployment of ML applications. One of the key components of SageMaker JumpStart is model hubs, which offer a vast catalog of pre-trained models, such as Mistral, for a variety of tasks.

    You can now discover and deploy Mercury and Mercury Coder in Amazon SageMaker Studio or programmatically through the SageMaker Python SDK, and derive model performance and MLOps controls with Amazon SageMaker AI features such as Amazon SageMaker Pipelines, Amazon SageMaker Debugger, or container logs. The model is deployed in a secure AWS environment and in your VPC, helping support data security for enterprise security needs.

    Prerequisites

    To deploy the Mercury models, make sure you have access to the recommended instance types based on the model size. To verify you have the necessary resources, complete the following steps:

    1. On the Service Quotas console, under AWS Services, choose Amazon SageMaker.
    2. Check that you have sufficient quota for the required instance type for endpoint deployment.
    3. Make sure at least one of these instance types is available in your target AWS Region.
    4. If needed, request a quota increase and contact your AWS account team for support.

    Make sure your SageMaker AWS Identity and Access Management (IAM) service role has the necessary permissions to deploy the model, including the following permissions to make AWS Marketplace subscriptions in the AWS account used:

    • aws-marketplace:ViewSubscriptions
    • aws-marketplace:Unsubscribe
    • aws-marketplace:Subscribe

    Alternatively, confirm your AWS account has a subscription to the model. If so, you can skip the following deployment instructions and start with subscribing to the model package.

    Subscribe to the model package

    To subscribe to the model package, complete the following steps:

    1. Open the model package listing page and choose Mercury or Mercury Coder.
    2. On the AWS Marketplace listing, choose Continue to subscribe.
    3. On the Subscribe to this software page, review and choose Accept Offer if you and your organization agree with the EULA, pricing, and support terms.
    4. Choose Continue to proceed with the configuration and then choose a Region where you have the service quota for the desired instance type.

    A product Amazon Resource Name (ARN) will be displayed. This is the model package ARN that you need to specify while creating a deployable model using Boto3.

    Deploy Mercury and Mercury Coder models on SageMaker JumpStart

    For those new to SageMaker JumpStart, you can use SageMaker Studio to access the Mercury and Mercury Coder models on SageMaker JumpStart.

    Deployment starts when you choose the Deploy option. You might be prompted to subscribe to this model through Amazon Bedrock Marketplace. If you are already subscribed, choose Deploy. After deployment is complete, you will see that an endpoint is created. You can test the endpoint by passing a sample inference request payload or by selecting the testing option using the SDK.

    Deploy Mercury using the SageMaker SDK

    In this section, we walk through deploying the Mercury model through the SageMaker SDK. You can follow a similar process for deploying the Mercury Coder model as well.

    To deploy the model using the SDK, copy the product ARN from the previous step and specify it in the model_package_arn in the following code:

    #Create the model package
    
    endpoint_name = name_from_base("mercury-endpoint")  # set this to your liking
    model = ModelPackage(role=role_arn, model_package_arn=package_arn, sagemaker_session=sagemaker_session)

    Deploy the model:

    # Deploy the Model. This may take 5-10 minutes to run
    
    instance_type = "ml.p5.48xlarge" # We only support ml.p5.48xlarge instances at the moment
    start = perf_counter()
    deployed_model = model.deploy(initial_instance_count=1, instance_type=instance_type, endpoint_name=endpoint_name)
    print(f"nDeployment took {perf_counter() - start:.2f} seconds")

    Use Mercury for code generation

    Let’s try asking the model to generate a simple tic-tac-toe game:

    payload = {
        "messages": [
           {
                "role": "user",
                "content": """
    Build a simple tic-tac-toe game.
    
    REQUIREMENTS:
    1. **Game**: 3x3 grid, human vs AI, click to play
    2. **AI**: Uses minimax to never lose (only win or draw)
    3. **Visualization**: Show AI's move scores in a simple list
    4. **Interface**: Grid + "New Game" button + move explanation
    
    IMPLEMENTATION:
    - Single HTML file with embedded CSS/JS
    - Basic minimax algorithm (no pruning needed)
    - Display: "AI chose position 5 (score: +10)" 
    - Clean, functional design
    
    DELIVERABLE:
    Working game that demonstrates perfect AI play with basic score visibility.
            """
            }
        ],
        "max_tokens": 2500,
    }
    start = perf_counter()
    outputs = predictor.predict(payload)
    eta = perf_counter() - start
    print(f"Speed: {outputs['usage']['completion_tokens'] / eta:.2f} tokens / secondn")
    print(outputs["choices"][0]["message"]["content"])

    We get the following response:

    Speed: 528.15 tokens / second
    
    ```html
    <!DOCTYPE html>
    <html>
    <head>
    <title>Tic-Tac-Toe with Unbeatable AI</title>
    <style>
      body {
        font-family: sans-serif;
        display: flex;
        flex-direction: column;
        align-items: center;
        justify-content: center;
        height: 100vh;
        margin: 0;
        background-color: #f0f0f0;
      }
    
      .grid {
        display: grid;
        grid-template-columns: repeat(3, 100px);
        gap: 10px;
        margin-bottom: 20px;
      }
    
      .cell {
        width: 100px;
        height: 100px;
        border: 1px solid #ccc;
        display: flex;
        align-items: center;
        justify-content: center;
        font-size: 2em;
        cursor: pointer;
        background-color: #fff;
      }
    
      .cell:hover {
        background-color: #e0e0e0;
      }
    
      .cell.x {
        color: red;
      }
    
      .cell.o {
        color: blue;
      }
    
      #aiMoves {
        margin-top: 20px;
        font-size: 1.2em;
      }
    
      #newGame {
        padding: 10px 20px;
        font-size: 1.2em;
        cursor: pointer;
        background-color: #4CAF50;
        color: white;
        border: none;
        border-radius: 5px;
      }
    
      #newGame:hover {
        background-color: #3e8e41;
      }
    </style>
    </head>
    <body>
    
    <h1>Tic-Tac-Toe</h1>
    
    <div class="grid" id="gameGrid">
      <div class="cell" data-index="0"></div>
      <div class="cell" data-index="1"></div>
      <div class="cell" data-index="2"></div>
      <div class="cell" data-index="3"></div>
      <div class="cell" data-index="4"></div>
      <div class="cell" data-index="5"></div>
      <div class="cell" data-index="6"></div>
      <div class="cell" data-index="7"></div>
      <div class="cell" data-index="8"></div>
    </div>
    
    <button id="newGame">New Game</button>
    
    <div id="aiMoves"></div>
    
    <script>
      const grid = document.getElementById('gameGrid');
      const cells = document.querySelectorAll('.cell');
      const aiMovesDiv = document.getElementById('aiMoves');
      const newGameButton = document.getElementById('newGame');
    
      let board = Array(9).fill(null);
      let currentPlayer = 'x';
      let gameActive = true;
      let aiMoveScores = [];
    
      function checkWin(player) {
        const winConditions = [
          [0, 1, 2], [3, 4, 5], [6, 7, 8], // Rows
          [0, 3, 6], [1, 4, 7], [2, 5, 8], // Columns
          [0, 4, 8], [2, 4, 6] // Diagonals
        ];
    
        return winConditions.some(condition => {
          return condition.every(index => board[index] === player);
        });
      }
    
      function checkDraw() {
        return board.every(cell => cell !== null);
      }
    
      function switchPlayer() {
        currentPlayer = currentPlayer === 'x' ? 'o' : 'x';
      }
    
      function updateBoard() {
        cells.forEach((cell, index) => {
          cell.textContent = board[index];
          cell.classList.remove('x', 'o');
          if (board[index] === 'x') {
            cell.classList.add('x');
          } else if (board[index] === 'o') {
            cell.classList.add('o');
          }
        });
      }
    
      function minimax(board, player) {
        let opponent = player === 'x' ? 'o' : 'x';
    
        if (checkWin(player)) {
          return { score: 10 };
        } else if (checkWin(opponent)) {
          return { score: -10 };
        } else if (checkDraw()) {
          return { score: 0 };
        }
    
        let moves = [];
        for (let i = 0; i < board.length; i++) {
          if (board[i] === null) {
            board[i] = player;
            let result = minimax(board, opponent);
            moves.push({ index: i, score: result.score });
            board[i] = null; // Undo move
          }
        }
    
        if (player === 'x') {
          let bestScore = -Infinity;
          let bestMove = null;
          for (let i = 0; i < moves.length; i++) {
            if (moves[i].score > bestScore) {
              bestScore = moves[i].score;
              bestMove = moves[i];
            }
          }
          return bestMove;
        } else {
          let bestScore = Infinity;
          let bestMove = null;
          for (let i = 0; i < moves.length; i++) {
            if (moves[i].score < bestScore) {
              bestScore = moves[i].score;
              bestMove = moves[i];
            }
          }
          return bestMove;
        }
      }
    
      function aiMove() {
        if (!gameActive) return;
    
        let bestMove = minimax(board, 'o');
        board[bestMove.index] = 'o';
        updateBoard();
        aiMoveScores.push(`AI chose position ${bestMove.index} (score: ${bestMove.score})`);
        aiMovesDiv.textContent = aiMoveScores.join('n');
    
        if (checkWin('o')) {
          alert('AI wins!');
          gameActive = false;
        } else if (checkDraw()) {
          alert('It's a draw!');
          gameActive = false;
        } else {
          currentPlayer = 'x';
        }
      }
    
      function handleCellClick(event) {
        const cellIndex = parseInt(event.target.dataset.index);
    
        if (gameActive && board[cellIndex] === null && currentPlayer === 'x') {
          board[cellIndex] = 'x';
          updateBoard();
    
          if (checkWin('x')) {
            alert('You win!');
            gameActive = false;
          } else if (checkDraw()) {
            alert('It's a draw!');
            gameActive = false;
          } else {
            aiMove();
          }
        }
      }
    
      function newGame() {
        board = Array(9).fill(null);
        currentPlayer = 'x';
        gameActive = true;
        aiMoveScores = [];
        aiMovesDiv.textContent = '';
        updateBoard();
      }
    
      cells.forEach(cell => cell.addEventListener('click', handleCellClick));
      newGameButton.addEventListener('click', newGame);
    </script>
    
    </body>
    </html>
    ```

    From the preceding response, we can see that the Mercury model generated a complete, functional tic-tac-toe game with minimax AI implementation at 528 tokens per second, delivering working HTML, CSS, and JavaScript in a single response. The code includes proper game logic, an unbeatable AI algorithm, and a clean UI with the specified requirements correctly implemented. This demonstrates strong code generation capabilities with exceptional speed for a diffusion-based model.

    Use Mercury for tool use and function calling

    Mercury models support advanced tool use capabilities, enabling them to intelligently determine when and how to call external functions based on user queries. This makes them ideal for building AI agents and assistants that can interact with external systems, APIs, and databases.

    Let’s demonstrate Mercury’s tool use capabilities by creating a travel planning assistant that can check weather and perform calculations:

    # Define available tools for the assistant
    tools = [
        {
            "type": "function",
            "function": {
                "name": "get_weather",
                "description": "Get the current weather in a given location",
                "parameters": {
                    "type": "object",
                    "properties": {
                        "location": {
                            "type": "string",
                            "description": "The city and state, e.g. San Francisco, CA"
                        },
                        "unit": {
                            "type": "string",
                            "enum": ["celsius", "fahrenheit"],
                            "description": "The unit of temperature"
                        }
                    },
                    "required": ["location"]
                }
            }
        },
        {
            "type": "function",
            "function": {
                "name": "calculate",
                "description": "Perform mathematical calculations",
                "parameters": {
                    "type": "object",
                    "properties": {
                        "expression": {
                            "type": "string",
                            "description": "The mathematical expression to evaluate"
                        }
                    },
                    "required": ["expression"]
                }
            }
        }
    ]
    #Create a travel planning query that requires multiple tools
    payload = {
        "messages": [
            {
                "role": "user",
                "content": "I'm planning a trip to Tokyo. Can you check the weather there and also tell me what 1000 USD is in Japanese Yen (use 1 USD = 150 JPY for calculation)?"
            }
        ],
        "tools": tools,
        "tool_choice": "auto",  # Let the model decide which tools to use
        "max_tokens": 2000,
        "temperature": 0.15
    }
    # Invoke the endpoint
    start = perf_counter()
    response = predictor.predict(payload)
    eta = perf_counter() - start
    # Display the tool calls requested by the model
    if 'choices' in response:
        message = response['choices'][0].get('message', {})
        if 'tool_calls' in message:
            print(f"Speed: {response['usage']['completion_tokens'] / eta:.2f} tokens/secondn")
            print(f"Mercury requested {len(message['tool_calls'])} tool calls:n")
        
            for i, tool_call in enumerate(message['tool_calls'], 1):
                func = tool_call.get('function', {})
                tool_name = func.get('name')
                args = json.loads(func.get('arguments', '{}'))
                
                print(f"Tool Call {i}:")
                print(f"  Function: {tool_name}")
                print(f"  Arguments: {json.dumps(args, indent=4)}")
                print()

    Expected response:

    Speed: 892.34 tokens/second
    Mercury requested 2 tool calls:
    Tool Call 1:
      Function: get_weather
      Arguments: {
        "location": "Tokyo, Japan",
        "unit": "celsius"
      }
    Tool Call 2:
      Function: calculate
      Arguments: {
        "expression": "1000 * 150"
      }

    After receiving the tool results, you can continue the conversation to get a natural language response:

    # Simulate tool execution results
    tool_results = [
        {
            "role": "tool",
            "tool_call_id": message['tool_calls'][0]['id'],
            "content": "The weather in Tokyo, Japan is 18°C and partly cloudy with a chance of rain."
        },
        {
            "role": "tool", 
            "tool_call_id": message['tool_calls'][1]['id'],
            "content": "The result is: 150000"
        }
    ]
    # Continue the conversation with tool results
    messages_with_results = [
        {"role": "user", "content": "I'm planning a trip to Tokyo. Can you check the weather there and also tell me what 1000 USD is in Japanese Yen (use 1 USD = 150 JPY for calculation)?"},
        message,  # Assistant's message with tool calls
        *tool_results  # Tool execution results
    ]
    final_payload = {
        "messages": messages_with_results,
        "max_tokens": 500
    }
    final_response = predictor.predict(final_payload)
    print(final_response['choices'][0]['message']['content'])

    Expected response:

    Based on the information I've gathered for your Tokyo trip:
    **Weather in Tokyo:**
    Currently, Tokyo is experiencing mild weather at 18°C (64°F) with partly cloudy skies and a chance of rain. I'd recommend bringing a light jacket and an umbrella just in case.
    **Currency Conversion:**
    1,000 USD converts to 150,000 Japanese Yen at the rate you specified (1 USD = 150 JPY). This should give you a good amount for expenses like meals, transportation, and shopping in Tokyo.
    For your trip planning, the mild temperature is perfect for sightseeing, though you'll want to have rain gear handy. The weather is comfortable for walking around popular areas like Shibuya, Shinjuku, or exploring temples and gardens.

    Clean up

    To avoid unwanted charges, complete the steps in this section to clean up your resources.

    Delete the Amazon Bedrock Marketplace deployment

    If you deployed the model using Amazon Bedrock Marketplace, complete the following steps:

    1. On the Amazon Bedrock console, in the navigation pane, under Foundation models, choose Marketplace deployments.
    2. Select the endpoint you want to delete, and on the Actions menu, choose Delete.
    3. Verify the endpoint details to make sure you’re deleting the correct deployment:
      1. Endpoint name
      2. Model name
      3. Endpoint status
    4. Choose Delete to delete the endpoint.
    5. In the Delete endpoint confirmation dialog, review the warning message, enter confirm, and choose Delete to permanently remove the endpoint.

    Delete the SageMaker JumpStart endpoint

    The SageMaker JumpStart model you deployed will incur costs if you leave it running. Use the following code to delete the endpoint if you want to stop incurring charges. For more details, see Delete Endpoints and Resources.

    sm.delete_model(ModelName=sm_model_name)
    sm.delete_endpoint_config(EndpointConfigName=endpoint_config_name)
    sm.delete_endpoint(EndpointName=endpoint_name)

    Conclusion

    In this post, we explored how you can access and deploy Mercury models using Amazon Bedrock Marketplace and SageMaker JumpStart. With support for both Mini and Small parameter sizes, you can choose the optimal model size for your specific use case. Visit SageMaker JumpStart in SageMaker Studio or Amazon Bedrock Marketplace to get started. For more information, refer to Use Amazon Bedrock tooling with Amazon SageMaker JumpStart models, Amazon SageMaker JumpStart Foundation Models, Getting started with Amazon SageMaker JumpStart, Amazon Bedrock Marketplace, and SageMaker JumpStart pretrained models.

    The Mercury family of diffusion-based large language models offers exceptional speed and performance, making it a powerful choice for your generative AI workloads with latency-sensitive requirements.


    About the authors

    Niithiyn Vijeaswaran is a Generative AI Specialist Solutions Architect with the Third-Party Model Science team at AWS. His area of focus is AWS AI accelerators (AWS Neuron). He holds a Bachelor’s degree in Computer Science and Bioinformatics.

    John Liu has 15 years of experience as a product executive and 9 years of experience as a portfolio manager. At AWS, John is a Principal Product Manager for Amazon Bedrock. Previously, he was the Head of Product for AWS Web3 / Blockchain. Prior to AWS, John held various product leadership roles at public blockchain protocols, fintech companies and also spent 9 years as a portfolio manager at various hedge funds.

    Jonathan Evans is a Worldwide Solutions Architect for Generative AI at AWS, where he helps customers leverage cutting-edge AI technologies with Anthropic’s Claude models on Amazon Bedrock, to solve complex business challenges. With a background in AI/ML engineering and hands-on experience supporting machine learning workflows in the cloud, Jonathan is passionate about making advanced AI accessible and impactful for organizations of all sizes.

    Rohit Talluri is a Generative AI GTM Specialist at Amazon Web Services (AWS). He is partnering with top generative AI model builders, strategic customers, key AI/ML partners, and AWS Service Teams to enable the next generation of artificial intelligence, machine learning, and accelerated computing on AWS. He was previously an Enterprise Solutions Architect and the Global Solutions Lead for AWS Mergers & Acquisitions Advisory.

    Breanne Warner is an Enterprise Solutions Architect at Amazon Web Services supporting healthcare and life science (HCLS) customers. She is passionate about supporting customers to use generative AI on AWS and evangelizing model adoption for first- and third-party models. Breanne is also Vice President of the Women at Amazon board with the goal of fostering inclusive and diverse culture at Amazon. Breanne holds a Bachelor’s of Science in Computer Engineering from the University of Illinois Urbana-Champaign.

    Source: Read More 

    Facebook Twitter Reddit Email Copy Link
    Previous ArticleUse Arduinos for Computer Vision
    Next Article How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    Related Posts

    Machine Learning

    How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark

    August 28, 2025
    Machine Learning

    Learn how Amazon Health Services improved discovery in Amazon search using AWS ML and gen AI

    August 27, 2025
    Leave A Reply Cancel Reply

    For security, use of Google's reCAPTCHA service is required which is subject to the Google Privacy Policy and Terms of Use.

    Continue Reading

    Postman introduces Agent Mode to integrate the power of AI agents into Postman’s core capabilities

    Tech & Work

    Summer Game Fest had a bit of a “weird” vibe this year — an extremely mixed bag of weak presentations and interesting titles

    News & Updates

    North Korean Hackers Flood npm Registry with XORIndex Malware in Ongoing Attack Campaign

    Development

    CVE-2025-7798 – Beijing Shenzhou Shihan Technology Multimedia Integrated Business Display System SQL Injection Vulnerability

    Common Vulnerabilities and Exposures (CVEs)

    Highlights

    How to screen record on your iPhone – it’s easy

    May 29, 2025

    You can easily capture a video of everything happening on your iPhone screen, including audio,…

    CVE-2025-5997 – Beamsec PhishPro Privileged API Abuse

    July 28, 2025

    CVE-2024-51360 – Apache Server PHP RCE

    May 23, 2025

    CVE-2025-50181 – “Open Redirect Vulnerability in urllib3”

    June 18, 2025
    © DevStackTips 2025. All rights reserved.
    • Contact
    • Privacy Policy

    Type above and press Enter to search. Press Esc to cancel.