How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark
When we began studying jailbreak evaluations, we found a fascinating paper claiming that you could jailbreak frontier LLMs simply by […]
When we began studying jailbreak evaluations, we found a fascinating paper claiming that you could jailbreak frontier LLMs simply by […]
Today, Amazon SageMaker HyperPod is announcing a new one-click, validated cluster creation experience that accelerates setup and prevents common misconfigurations,
Intelligent document processing (IDP) refers to the automated extraction, classification, and processing of data from various document formats—both structured and
Retrieval Augmented Generation (RAG) is a powerful approach for building generative AI applications by providing foundation models (FMs) access to
In this post, we explore how natural language database analytics can revolutionize the way organizations interact with their structured data
As organizations increasingly adopt foundation models (FMs) for their artificial intelligence and machine learning (AI/ML) workloads, managing large-scale inference operations
Training and deploying large AI models requires advanced distributed computing capabilities, but managing these distributed systems shouldn’t be complex for
Today, we’re excited to announce that Amazon SageMaker HyperPod now supports managed node automatic scaling with Karpenter, so you can
When deploying AI agents to Amazon Bedrock AgentCore Runtime (currently in preview), customers often want to use custom domain names
This post was co-written with Nick Frichette and Vijay George from Datadog. As organizations increasingly adopt Amazon Bedrock for generative
Finance analysts across Amazon Finance face mounting complexity in financial planning and analysis processes. When working with vast datasets spanning
Air pollution remains one of Africa’s most pressing environmental health crises, causing widespread illness across the continent. Organizations like sensors.AFRICA
This post is co-written with Julieta Rappan, Macarena Blasi, and María Candela Blanco from the Government of the City of
Today, we are excited to announce that Mercury and Mercury Coder foundation models (FMs) from Inception Labs are available through
Healthcare discovery on ecommerce domains presents unique challenges that traditional product search wasn’t designed to handle. Unlike searching for books
Language models must be adapted to understand and follow user instructions. Reinforcement learning is widely used to facilitate this —
We introduce SlowFast-LLaVA-1.5 (abbreviated as SF-LLaVA-1.5), a family of video large language models (LLMs) offering a token-efficient solution for long-form
Amazon SageMaker HyperPod is a purpose-built infrastructure for optimizing foundation model (FM) training and inference at scale. SageMaker HyperPod removes
Every day, organizations process millions of documents, including invoices, contracts, insurance claims, medical records, and financial statements. Despite the critical
Most organizations evaluating foundation models limit their analysis to three primary dimensions: accuracy, latency, and cost. While these metrics provide