This site uses cookies to improve your experience. To help us insure we adhere to various privacy regulations, please select your country/region of residence. If you do not select a country, we will assume you are from the United States. Select your Cookie Settings or view our Privacy Policy and Terms of Use.
Cookie Settings
Cookies and similar technologies are used on this website for proper function of the website, for tracking performance analytics and for marketing purposes. We and some of our third-party providers may use cookie data for various purposes. Please review the cookie settings below and choose your preference.
Used for the proper function of the website
Used for monitoring website traffic and interactions
Cookie Settings
Cookies and similar technologies are used on this website for proper function of the website, for tracking performance analytics and for marketing purposes. We and some of our third-party providers may use cookie data for various purposes. Please review the cookie settings below and choose your preference.
Strictly Necessary: Used for the proper function of the website
Performance/Analytics: Used for monitoring website traffic and interactions
In this post, we guide you through integrating Amazon Bedrock Agents with enterprise data APIs to create more personalized and effective customer support experiences. Although the principles discussed are applicable across various industries, we use an automotive parts retailer as our primary example throughout this post.
Amazon Bedrock announces the preview launch of Session Management APIs, a new capability that enables developers to simplify state and context management for generative AI applications built with popular open source frameworks such as LangGraph and LlamaIndex. Building generative AI applications requires more than model API calls.
For example, given one phrasing of a question, the model can claim to not know the answer, but given a slight rephrase, can answer correctly.” Moreover, it does not offer handy out-of-the-box integrations to your CCaaS or CRM systems for example.
With this solution, you can interact directly with the chat assistant powered by AWS from your Google Chat environment, as shown in the following example. The custom Google Chat app, configured for HTTP integration, sends an HTTP request to an API Gateway endpoint. The following figure illustrates the high-level design of the solution.
In addition, they use the developer-provided instruction to create an orchestration plan and then carry out the plan by invoking company APIs and accessing knowledge bases using Retrieval Augmented Generation (RAG) to provide an answer to the users request. Lets consider our HR agent example.
As an example, climbing a wind turbine in bad weather for an inspection can be dangerous. The following figure shows an example of the user dashboard and drone conversation. The following figure is an example of drone 4K footage. Plus, even the best human inspector can miss things.
It also uses a number of other AWS services such as Amazon API Gateway , AWS Lambda , and Amazon SageMaker. It contains services used to onboard, manage, and operate the environment, for example, to onboard and off-board tenants, users, and models, assign quotas to different tenants, and authentication and authorization microservices.
One can quickly host such application on the AWS Cloud without managing the underlying infrastructure, for example, with Amazon Simple Storage Service (S3) and Amazon CloudFront. Note that these APIs use objects as namespaces, alleviating the need for explicit imports. Here, we use Anthropics Claude 3.5 Sonnet).
By using the power of LLMs and combining them with specialized tools and APIs, agents can tackle complex, multistep tasks that were previously beyond the reach of traditional AI systems. Whenever local database information is unavailable, it triggers an online search using the Tavily API. Its used by the weather_agent() function.
Clone the repo To get started, clone the repository by running the following command, and then switch to the working directory: git clone [link] Build your guardrail To build the guardrail, you can use the CreateGuardrail API. For example, a banking assistant application can be designed to deny topics related to illegal investment advice.
We walk through the key components and services needed to build the end-to-end architecture, offering example code snippets and explanations for each critical element that help achieve the core functionality. With Lambda integration, we can create a web API with an endpoint to the Lambda function.
For example, by the end of this tutorial, you will be able to query the data with prompts such as “Can you return our five top selling products this quarter and the principal customer complaints for each?” This includes setting up Amazon API Gateway , AWS Lambda functions, and Amazon Athena to enable querying the structured sales data.
Amazon Bedrock Flows offers an intuitive visual builder and a set of APIs to seamlessly link foundation models (FMs), Amazon Bedrock features, and AWS services to build and automate user-defined generative AI workflows at scale. For our example, we chose Amazons Nova Lite model and set the temperature inference parameter to 0.1
Traditional automation approaches require custom API integrations for each application, creating significant development overhead. For example, your agent could take screenshots, create and edit text files, and run built-in Linux commands. The output is given back to the Amazon Bedrock agent for further processing.
The new ApplyGuardrail API enables you to assess any text using your preconfigured guardrails in Amazon Bedrock, without invoking the FMs. In this post, we demonstrate how to use the ApplyGuardrail API with long-context inputs and streaming outputs. For example, you can now use the API with models hosted on Amazon SageMaker.
Then we provide examples of how to use the AI-powered chat interface to gain insights from the connected data source. We provide the service account with authorization scopes to allow access to the required Gmail APIs. In our example, we name the project GmailConnector. Choose Enable to enable this API. Choose Create.
This involves creating an OAuth API endpoint in ServiceNow and using the web experience URL from Amazon Q Business as the callback URL. The final step of the solution involves enhancing the application environment with a custom plugin for ServiceNow using APIs defined in an OpenAPI schema.
Customers can use the SageMaker Studio UI or APIs to specify the SageMaker Model Registry model to be shared and grant access to specific AWS accounts or to everyone in the organization. With this launch, customers can now seamlessly share and access ML models registered in SageMaker Model Registry between different AWS accounts.
Using SageMaker with MLflow to track experiments The fully managed MLflow capability on SageMaker is built around three core components: MLflow tracking server This component can be quickly set up through the Amazon SageMaker Studio interface or using the API for more granular configurations.
This could be APIs, code functions, or schemas and structures required by your end application. In this post, we discuss tool use and the new tool choice feature, with example use cases. For example, if a user asks What is the weather in Seattle? For example, if a user asks What is the weather in Seattle?
Amazon Bedrock is a fully managed service that offers a choice of high-performing foundation models (FMs) from leading AI companies like AI21 Labs, Anthropic, Cohere, Meta, Mistral AI, Stability AI, and Amazon through a single API, along with a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI.
Intricate workflows that require dynamic and complex API orchestration can often be complex to manage. In this post, we explore how chaining domain-specific agents using Amazon Bedrock Agents can transform a system of complex API interactions into streamlined, adaptive workflows, empowering your business to operate with agility and precision.
You can get started without any prior machine learning (ML) experience, and Amazon Personalize allows you to use APIs to build sophisticated personalization capabilities. In our example, we use the Top picks for you recommender. In this example, we create a Top picks for you recommender. Create a recommender.
adds new APIs to customize GraphStorm pipelines: you now only need 12 lines of code to implement a custom node classification training loop. To help you get started with the new API, we have published two Jupyter notebook examples: one for node classification, and one for a link prediction task. Specifically, GraphStorm 0.3
Reduced time and effort in testing and deploying AI workflows with SDK APIs and serverless infrastructure. We can also quickly integrate flows with our applications using the SDK APIs for serverless flow execution — without wasting time in deployment and infrastructure management. For example, CustomerServiceGuardrail-001.
In this post, we show you an example of a generative AI assistant application and demonstrate how to assess its security posture using the OWASP Top 10 for Large Language Model Applications , as well as how to apply mitigations for common threats. These steps might involve both the use of an LLM and external data sources and APIs.
Then we deep dive into the new rolling update feature for inference components and provide practical examples using DeepSeek distilled models to demonstrate this feature. Consider an example where a customer has 10 copies of an inference component spread across 5 ml.p4d.24xlarge You can find the example notebook in the GitHub repo.
We provide practical examples for both SCP modifications and AWS Control Tower implementations. Importantly, cross-Region inference prioritizes the connected Amazon Bedrock API source Region when possible, helping minimize latency and improve overall responsiveness. This completes the configuration.
We also showcase a real-world example for predicting the root cause category for support cases. For the use case of labeling the support root cause categories, its often harder to source examples for categories such as Software Defect, Feature Request, and Documentation Improvement for labeling than it is for Customer Education.
The following table provides example questions with their domain and question type. Amazon Bedrock APIs make it straightforward to use Amazon Titan Text Embeddings V2 for embedding data. The eight different question types are simple , simple_w_condition , comparison , aggregation , set , false_premise , post-processing , and multi-hop.
Beyond Amazon Bedrock models, the service offers the flexible ApplyGuardrails API that enables you to assess text using your pre-configured guardrails without invoking FMs, allowing you to implement safety controls across generative AI applicationswhether running on Amazon Bedrock or on other systemsat both input and output levels.
This serves as an example of how generative AI can streamline operations that involve diverse data types and formats. The solution uses the FMs tool use capabilities, accessed through the Amazon Bedrock Converse API. Use case and dataset For our example use case, we examine a patient intake process at a healthcare institution.
Amazon Bedrock is a fully managed service that offers a choice of high-performing foundation models (FMs) from leading AI companies like AI21 Labs, Anthropic, Cohere, Meta, Mistral AI, Stability AI, and Amazon through a single API, along with a broad set of capabilities to build generative AI applications with security, privacy, and responsible AI.
For enterprise data, a major difficulty stems from the common case of database tables having embedded structures that require specific knowledge or highly nuanced processing (for example, an embedded XML formatted string). As a result, NL2SQL solutions for enterprise data are often incomplete or inaccurate.
Cloud providers have recognized the need to offer model inference through an API call, significantly streamlining the implementation of AI within applications. Although a single API call can address simple use cases, more complex ones may necessitate the use of multiple calls and integrations with other services.
See the following figure for an example. The following example illustrates the hybrid RAG high-level architecture. The embedding model, which is hosted on the same EC2 instance as the local LLM API inference server, converts the text chunks into vector representations.
24xlarge", hyperparameters=my_hyperparameters, ) estimator.fit({"training": train_data_location}) For a custom model, see the example notebooks in GitHub. Update models in the private hub Modify your existing private HubContent by calling the new sagemaker:UpdateHubContent API. Refer to the public API documentation for more details.
Solution overview The following code is an example metadata filter for Amazon Bedrock Knowledge Bases. We have provided example documents and metadata in the accompanying GitHub repo for you to upload. This example data contains user answers to an online questionnaire about travel preferences.
By choosing View API , you can also access the model using code examples in the AWS Command Line Interface (AWS CLI) and AWS SDKs. Pixtral Large use cases In this section, we provide example use cases of Pixtral Large using sample prompts. Additionally, Pixtral Large supports the Converse API and tool usage.
Amazon Bedrock is a fully managed service that offers a choice of high-performing foundation models (FMs) from leading AI companies such as AI21 Labs, Anthropic, Cohere, Meta, Stability AI, and Amazon through a single API, along with a broad set of capabilities you need to build generative AI applications with security, privacy, and responsible AI.
Whether youre new to AI development or an experienced practitioner, this post provides step-by-step guidance and code examples to help you build more reliable AI applications. Lets walkthrough an example of how this solution would handle a users question. For example, if the question was What hotels are near re:Invent?
The organizations that figure this out first will have a significant competitive advantageand were already seeing compelling examples of whats possible. The question is no longer whether to adopt generative AI, but how to move from promising pilots to production-ready systems that deliver real business value.
Enabling Global Resiliency for an Amazon Lex bot is straightforward using the AWS Management Console , AWS Command Line Interface (AWS CLI), or APIs. For this example, we create a bot named BookHotel in the source Region ( us-east-1 ). Global Resiliency APIs Global Resiliency provides API support to create and manage replicas.
You can retrieve the number of copies of an inference component at any time by making the DescribeInferenceComponent API call and checking the CurrentCopyCount. In the following code example, we set the TargetValue to 5. The specific permissions needed depend on the target API being called. import json scheduler = boto3.client('scheduler')
We organize all of the trending information in your field so you don't have to. Join 34,000+ users and stay up to date on the latest articles your peers are reading.
You know about us, now we want to get to know you!
Let's personalize your content
Let's get even more personalized
We recognize your account from another site in our network, please click 'Send Email' below to continue with verifying your account and setting a password.
Let's personalize your content