Enhancing LLM accuracy with Coveo Passage Retrieval on Amazon Bedrock

Favorite This post is co-written with Keith Beaudoin and Nicolas Bordeleau from Coveo. As generative AI transforms business operations, enterprises face a critical challenge: how can they help large language models (LLMs) provide accurate and trustworthy responses? Without reliable data foundations, these AI models can generate misleading or inaccurate responses,

Read More
Shared by AWS Machine Learning September 4, 2025

Unlocking the future of professional services: How Proofpoint uses Amazon Q Business

Favorite This post was written with Stephen Coverdale and Alessandra Filice of Proofpoint. At the forefront of cybersecurity innovation, Proofpoint has redefined its professional services by integrating Amazon Q Business, a fully managed, generative AI powered assistant that you can configure to answer questions, provide summaries, generate content, and complete

Read More
Shared by AWS Machine Learning September 4, 2025

Announcing the new cluster creation experience for Amazon SageMaker HyperPod

Favorite Today, Amazon SageMaker HyperPod is announcing a new one-click, validated cluster creation experience that accelerates setup and prevents common misconfigurations, so you can launch your distributed training and inference clusters complete with Slurm or Amazon Elastic Kubernetes Service (Amazon EKS) orchestration, Amazon Virtual Private Cloud (Amazon VPC) networking, high-performance

Read More
Shared by AWS Machine Learning September 3, 2025

Document intelligence evolved: Building and evaluating KIE solutions that scale

Favorite Intelligent document processing (IDP) refers to the automated extraction, classification, and processing of data from various document formats—both structured and unstructured. Within the IDP landscape, key information extraction (KIE) serves as a fundamental component, enabling systems to identify and extract critical data points from documents with minimal human intervention.

Read More
Shared by AWS Machine Learning September 3, 2025

Deploy Amazon Bedrock Knowledge Bases using Terraform for RAG-based generative AI applications

Favorite Retrieval Augmented Generation (RAG) is a powerful approach for building generative AI applications by providing foundation models (FMs) access to additional, relevant data. This approach improves response accuracy and transparency while avoiding the potential cost and complexity of FM training or fine-tuning. Many customers use Amazon Bedrock Knowledge Bases

Read More
Shared by AWS Machine Learning September 3, 2025

Natural language-based database analytics with Amazon Nova

Favorite In this post, we explore how natural language database analytics can revolutionize the way organizations interact with their structured data through the power of large language model (LLM) agents. Natural language interfaces to databases have long been a goal in data management. Agents enhance database analytics by breaking down

Read More
Shared by AWS Machine Learning September 3, 2025

Build a serverless Amazon Bedrock batch job orchestration workflow using AWS Step Functions

Favorite As organizations increasingly adopt foundation models (FMs) for their artificial intelligence and machine learning (AI/ML) workloads, managing large-scale inference operations efficiently becomes crucial. Amazon Bedrock supports two general types of large-scale inference patterns: real-time inference and batch inference for use cases that involve processing massive datasets where immediate results

Read More
Shared by AWS Machine Learning September 3, 2025

Train and deploy models on Amazon SageMaker HyperPod using the new HyperPod CLI and SDK

Favorite Training and deploying large AI models requires advanced distributed computing capabilities, but managing these distributed systems shouldn’t be complex for data scientists and machine learning (ML) practitioners. The newly released command line interface (CLI) and software development kit (SDK) for Amazon SageMaker HyperPod simplify how you can use the

Read More
Shared by AWS Machine Learning September 3, 2025

Introducing auto scaling on Amazon SageMaker HyperPod

Favorite Today, we’re excited to announce that Amazon SageMaker HyperPod now supports managed node automatic scaling with Karpenter, so you can efficiently scale your SageMaker HyperPod clusters to meet your inference and training demands. Real-time inference workloads require automatic scaling to address unpredictable traffic patterns and maintain service level agreements

Read More
Shared by AWS Machine Learning August 30, 2025