AWS Machine Learning Blog

Artificial Intelligence Official Machine Learning Blog of Amazon Web Services

  • How Myriad Genetics achieved fast, accurate, and cost-efficient document processing using the AWS open-source Generative AI Intelligent Document Processing Accelerator
    by Priyashree Roy on November 27, 2025 at 12:58 am

    In this post, we explore how Myriad Genetics partnered with the AWS Generative AI Innovation Center to transform their healthcare document processing pipeline using Amazon Bedrock and Amazon Nova foundation models, achieving 98% classification accuracy while reducing costs by 77% and processing time by 80%. We detail the technical implementation using AWS’s open-source GenAI Intelligent Document Processing Accelerator, the optimization strategies for document classification and key information extraction, and the measurable business impact on Myriad’s prior authorization workflows.

  • How CBRE powers unified property management search and digital assistant using Amazon Bedrock
    by Lokesha Thimmegowda, Muppirala Venkata Krishna Kumar, Maraka Vishwadev on November 27, 2025 at 12:56 am

    In this post, CBRE and AWS demonstrate how they transformed property management by building a unified search and digital assistant using Amazon Bedrock, enabling professionals to access millions of documents and multiple databases through natural language queries. The solution combines Amazon Nova Pro for SQL generation and Claude Haiku for document interactions, achieving a 67% reduction in processing time while maintaining enterprise-grade security across more than eight million documents.

  • Managed Tiered KV Cache and Intelligent Routing for Amazon SageMaker HyperPod
    by Chaitanya Hazarey on November 27, 2025 at 12:50 am

    In this post, we introduce Managed Tiered KV Cache and Intelligent Routing for Amazon SageMaker HyperPod, new capabilities that can reduce time to first token by up to 40% and lower compute costs by up to 25% for long context prompts and multi-turn conversations. These features automatically manage distributed KV caching infrastructure and intelligent request routing, making it easier to deploy production-scale LLM inference workloads with enterprise-grade performance while significantly reducing operational overhead.

  • Apply fine-grained access control with Bedrock AgentCore Gateway interceptors
    by Dhawalkumar Patel on November 26, 2025 at 10:28 pm

    We are launching a new feature: gateway interceptors for Amazon Bedrock AgentCore Gateway. This powerful new capability provides fine-grained security, dynamic access control, and flexible schema management.

  • How Condé Nast accelerated contract processing and rights analysis with Amazon Bedrock
    by Bob Boiko, Christopher Donnellan, Sarat Tatavarthi on November 26, 2025 at 9:37 pm

    In this post, we explore how Condé Nast used Amazon Bedrock and Anthropic’s Claude to accelerate their contract processing and rights analysis workstreams. The company’s extensive portfolio, spanning multiple brands and geographies, required managing an increasingly complex web of contracts, rights, and licensing agreements.

  • Building AI-Powered Voice Applications: Amazon Nova Sonic Telephony Integration Guide
    by Reilly Manton on November 26, 2025 at 9:21 pm

    Available through the Amazon Bedrock bidirectional streaming API, Amazon Nova Sonic can connect to your business data and external tools and can be integrated directly with telephony systems. This post will introduce sample implementations for the most common telephony scenarios.

  • University of California Los Angeles delivers an immersive theater experience with AWS generative AI services
    by Aditya Singh on November 26, 2025 at 9:20 pm

    In this post, we will walk through the performance constraints and design choices by OARC and REMAP teams at UCLA, including how AWS serverless infrastructure, AWS Managed Services, and generative AI services supported the rapid design and deployment of our solution. We will also describe our use of Amazon SageMaker AI and how it can be used reliably in immersive live experiences.

  • Optimizing Mobileye’s REM™ with AWS Graviton: A focus on ML inference and Triton integration
    by Chaim Rand, Pini Reisman, and Eliyah Weinberg on November 26, 2025 at 7:50 pm

    In this post, we focus on one portion of the REM™ system: the automatic identification of changes to the road structure which we will refer to as Change Detection. We will share our journey of architecting and deploying a solution for Change Detection, the core of which is a deep learning model called CDNet. We will share real-life decisions and tradeoffs when building and deploying a high-scale, highly parallelized algorithmic pipeline based on a Deep Learning (DL) model, with an emphasis on efficiency and throughput.

  • Evaluate models with the Amazon Nova evaluation container using Amazon SageMaker AI
    by Tony Santiago on November 26, 2025 at 7:39 pm

    This blog post introduces the new Amazon Nova model evaluation features in Amazon SageMaker AI. This release adds custom metrics support, LLM-based preference testing, log probability capture, metadata analysis, and multi-node scaling for large evaluations.

  • Beyond the technology: Workforce changes for AI
    by Taimur Rashid on November 26, 2025 at 6:42 pm

    In this post, we explore three essential strategies for successfully integrating AI into your organization: addressing organizational debt before it compounds, embracing distributed decision-making through the “octopus organization” model, and redefining management roles to align with AI-powered workflows. Organizations must invest in both technology and workforce preparation, focusing on streamlining processes, empowering teams with autonomous decision-making within defined parameters, and evolving each management layer from traditional oversight to mentorship, quality assurance, and strategic vision-setting.

  • Enhanced performance for Amazon Bedrock Custom Model Import
    by Nick McCarthy on November 26, 2025 at 4:46 pm

    You can now achieve significant performance improvements when using Amazon Bedrock Custom Model Import, with reduced end-to-end latency, faster time-to-first-token, and improved throughput through advanced PyTorch compilation and CUDA graph optimizations. With Amazon Bedrock Custom Model Import you can to bring your own foundation models to Amazon Bedrock for deployment and inference at scale. In this post, we introduce how to use the improvements in Amazon Bedrock Custom Model Import.

  • Amazon SageMaker AI introduces EAGLE based adaptive speculative decoding to accelerate generative AI inference
    by Kareem Syed-Mohammed on November 26, 2025 at 12:29 am

    Amazon SageMaker AI now supports EAGLE-based adaptive speculative decoding, a technique that accelerates large language model inference by up to 2.5x while maintaining output quality. In this post, we explain how to use EAGLE 2 and EAGLE 3 speculative decoding in Amazon SageMaker AI, covering the solution architecture, optimization workflows using your own datasets or SageMaker’s built-in data, and benchmark results demonstrating significant improvements in throughput and latency.

  • Train custom computer vision defect detection model using Amazon SageMaker
    by Ryan Vanderwerf on November 25, 2025 at 10:44 pm

    In this post, we demonstrate how to migrate computer vision workloads from Amazon Lookout for Vision to Amazon SageMaker AI by training custom defect detection models using pre-trained models available on AWS Marketplace. We provide step-by-step guidance on labeling datasets with SageMaker Ground Truth, training models with flexible hyperparameter configurations, and deploying them for real-time or batch inference—giving you greater control and flexibility for automated quality inspection use cases.

  • Practical implementation considerations to close the AI value gap
    by Bhargs Srivathsan on November 25, 2025 at 8:19 pm

    The AWS Customer Success Center of Excellence (CS COE) helps customers get tangible value from their AWS investments. We’ve seen a pattern: customers who build AI strategies that address people, process, and technology together succeed more often. In this post, we share practical considerations that can help close the AI value gap.

  • Introducing bidirectional streaming for real-time inference on Amazon SageMaker AI
    by Lingran Xia on November 25, 2025 at 7:09 pm

    We’re introducing bidirectional streaming for Amazon SageMaker AI Inference, which transforms inference from a transactional exchange into a continuous conversation. This post shows you how to build and deploy a container with bidirectional streaming capability to a SageMaker AI endpoint. We also demonstrate how you can bring your own container or use our partner Deepgram’s pre-built models and containers on SageMaker AI to enable bi-directional streaming feature for real-time inference.

  • Physical AI in practice: Technical foundations that fuel human-machine interactions
    by Sri Elaprolu, Alla Simoneau, Paul Amadeo, and Laura Kulowski on November 25, 2025 at 5:00 pm

    In this post, we explore the complete development lifecycle of physical AI—from data collection and model training to edge deployment—and examine how these intelligent systems learn to understand, reason, and interact with the physical world through continuous feedback loops. We illustrate this workflow through Diligent Robotics’ Moxi, a mobile manipulation robot that has completed over 1.2 million deliveries in hospitals, saving nearly 600,000 hours for clinical staff while transforming healthcare logistics and returning valuable time to patient care.

  • HyperPod now supports Multi-Instance GPU to maximize GPU utilization for generative AI tasks
    by Aman Shanbhag on November 25, 2025 at 4:10 pm

    In this post, we explore how Amazon SageMaker HyperPod now supports NVIDIA Multi-Instance GPU (MIG) technology, enabling you to partition powerful GPUs into multiple isolated instances for running concurrent workloads like inference, research, and interactive development. By maximizing GPU utilization and reducing wasted resources, MIG helps organizations optimize costs while maintaining performance isolation and predictable quality of service across diverse machine learning tasks.

  • Accelerate generative AI innovation in Canada with Amazon Bedrock cross-Region inference
    by Daniel Duplessis on November 24, 2025 at 11:56 pm

    We are excited to announce that customers in Canada can now access advanced foundation models including Anthropic’s Claude Sonnet 4.5 and Claude Haiku 4.5 on Amazon Bedrock through cross-Region inference (CRIS). This post explores how Canadian organizations can use cross-Region inference profiles from the Canada (Central) Region to access the latest foundation models to accelerate AI initiatives. We will demonstrate how to get started with these new capabilities, provide guidance for migrating from older models, and share recommended practices for quota management.

  • Power up your ML workflows with interactive IDEs on SageMaker HyperPod
    by Durga Sury on November 24, 2025 at 9:25 pm

    Amazon SageMaker HyperPod clusters with Amazon Elastic Kubernetes Service (EKS) orchestration now support creating and managing interactive development environments such as JupyterLab and open source Visual Studio Code, streamlining the ML development lifecycle by providing managed environments for familiar tools to data scientists. This post shows how HyperPod administrators can configure Spaces for their clusters, and how data scientists can create and connect to these Spaces.

  • Claude Opus 4.5 now in Amazon Bedrock
    by Jonathan Evans on November 24, 2025 at 7:22 pm

    Anthropic’s newest foundation model, Claude Opus 4.5, is now available in Amazon Bedrock, a fully managed service that offers a choice of high-performing foundation models from leading AI companies. In this post, I’ll show you what makes this model different, walk through key business applications, and demonstrate how to use Opus 4.5’s new tool use capabilities on Amazon Bedrock.

Share Websitecyber
We are an ethical website cyber security team and we perform security assessments to protect our clients.