Remove APIs Remove Metrics Remove Presentation
article thumbnail

Amazon Bedrock launches Session Management APIs for generative AI applications (Preview)

AWS Machine Learning

Amazon Bedrock announces the preview launch of Session Management APIs, a new capability that enables developers to simplify state and context management for generative AI applications built with popular open source frameworks such as LangGraph and LlamaIndex. Building generative AI applications requires more than model API calls.

APIs 119
article thumbnail

Build a multi-tenant generative AI environment for your enterprise on AWS

AWS Machine Learning

It also uses a number of other AWS services such as Amazon API Gateway , AWS Lambda , and Amazon SageMaker. API Gateway is serverless and hence automatically scales with traffic. API Gateway also provides a WebSocket API. As a result, building such a solution is often a significant undertaking for IT teams.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Centralize model governance with SageMaker Model Registry Resource Access Manager sharing

AWS Machine Learning

Customers can use the SageMaker Studio UI or APIs to specify the SageMaker Model Registry model to be shared and grant access to specific AWS accounts or to everyone in the organization. With this launch, customers can now seamlessly share and access ML models registered in SageMaker Model Registry between different AWS accounts.

article thumbnail

Build generative AI applications quickly with Amazon Bedrock IDE in Amazon SageMaker Unified Studio

AWS Machine Learning

Building generative AI applications presents significant challenges for organizations: they require specialized ML expertise, complex infrastructure management, and careful orchestration of multiple services. This includes setting up Amazon API Gateway , AWS Lambda functions, and Amazon Athena to enable querying the structured sales data.

APIs 106
article thumbnail

Considerations for addressing the core dimensions of responsible AI for Amazon Bedrock applications

AWS Machine Learning

The rapid advancement of generative AI promises transformative innovation, yet it also presents significant challenges. For automatic model evaluation jobs, you can either use built-in datasets across three predefined metrics (accuracy, robustness, toxicity) or bring your own datasets.

APIs 110
article thumbnail

Build a video insights and summarization engine using generative AI with Amazon Bedrock

AWS Machine Learning

This post presents a solution where you can upload a recording of your meeting (a feature available in most modern digital communication services such as Amazon Chime ) to a centralized video insights and summarization engine. All of this data is centralized and can be used to improve metrics in scenarios such as sales or call centers.

article thumbnail

Unlock cost savings with the new scale down to zero feature in SageMaker Inference

AWS Machine Learning

The scale down to zero feature presents new opportunities for how businesses can approach their cloud-based ML operations. Use faster auto scaling metrics – Take advantage of more granular auto scaling metrics like ConcurrentRequestsPerCopy to more accurately monitor and react to changes in inference traffic.

APIs 107