article thumbnail

Reducing hallucinations in LLM agents with a verified semantic cache using Amazon Bedrock Knowledge Bases

AWS Machine Learning

Solution overview Our solution implements a verified semantic cache using the Amazon Bedrock Knowledge Bases Retrieve API to reduce hallucinations in LLM responses while simultaneously improving latency and reducing costs. The function checks the semantic cache (Amazon Bedrock Knowledge Bases) using the Retrieve API.

article thumbnail

Use RAG for drug discovery with Knowledge Bases for Amazon Bedrock

AWS Machine Learning

Knowledge Bases for Amazon Bedrock allows you to build performant and customized Retrieval Augmented Generation (RAG) applications on top of AWS and third-party vector stores using both AWS and third-party models. RAG is a popular technique that combines the use of private data with large language models (LLMs).

APIs 137
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Build a self-service digital assistant using Amazon Lex and Knowledge Bases for Amazon Bedrock

AWS Machine Learning

Generative artificial intelligence (AI)-powered chatbots play a crucial role in delivering human-like interactions by providing responses from a knowledge base without the involvement of live agents. You can simply connect QnAIntent to company knowledge sources and the bot can immediately handle questions using the allowed content.

article thumbnail

Create a generative AI–powered custom Google Chat application using Amazon Bedrock

AWS Machine Learning

The Lambda function interacts with Amazon Bedrock through its runtime APIs, using either the RetrieveAndGenerate API that connects to a knowledge base, or the Converse API to chat directly with an LLM available on Amazon Bedrock. If you don’t have an existing knowledge base, refer to Create an Amazon Bedrock knowledge base.

APIs 124
article thumbnail

Transforming credit decisions using generative AI with Rich Data Co and AWS

AWS Machine Learning

To solve this challenge, RDC used generative AI , enabling teams to use its solution more effectively: Data science assistant Designed for data science teams, this agent assists teams in developing, building, and deploying AI models within a regulated environment.

article thumbnail

Revolutionizing clinical trials with the power of voice and AI

AWS Machine Learning

This transcription then serves as the input for a powerful LLM, which draws upon its vast knowledge base to provide personalized, context-aware responses tailored to your specific situation. These data sources provide contextual information and serve as a knowledge base for the LLM.

article thumbnail

How Formula 1® uses generative AI to accelerate race-day issue resolution

AWS Machine Learning

The transformed logs were stored in a separate S3 bucket, while another EventBridge schedule fed these transformed logs into Amazon Bedrock Knowledge Bases , an end-to-end managed Retrieval Augmented Generation (RAG) workflow capability, allowing the chat assistant to query them efficiently.

APIs 71