AWS Machine Learning Blog Official Machine Learning Blog of Amazon Web Services
- Deploy Qwen models with Amazon Bedrock Custom Model Importby Ajit Mahareddy on June 13, 2025 at 5:17 pm
You can now import custom weights for Qwen2, Qwen2_VL, and Qwen2_5_VL architectures, including models like Qwen 2, 2.5 Coder, Qwen 2.5 VL, and QwQ 32B. In this post, we cover how to deploy Qwen 2.5 models with Amazon Bedrock Custom Model Import, making them accessible to organizations looking to use state-of-the-art AI capabilities within the AWS infrastructure at an effective cost.
- Build generative AI solutions with Amazon Bedrockby Sajjan AVS on June 13, 2025 at 4:06 pm
In this post, we show you how to build generative AI applications on Amazon Web Services (AWS) using the capabilities of Amazon Bedrock, highlighting how Amazon Bedrock can be used at each step of your generative AI journey. This guide is valuable for both experienced AI engineers and newcomers to the generative AI space, helping you use Amazon Bedrock to its fullest potential.
- How Netsertive built a scalable AI assistant to extract meaningful insights from real-time data using Amazon Bedrock and Amazon Novaby Nicholas Switzer on June 13, 2025 at 4:03 pm
In this post, we show how Netsertive introduced a generative AI-powered assistant into MLX, using Amazon Bedrock and Amazon Nova, to bring their next generation of the platform to life.
- Make videos accessible with automated audio descriptions using Amazon Novaby Dylan Martin on June 13, 2025 at 4:01 pm
In this post, we demonstrate how you can use services like Amazon Nova, Amazon Rekognition, and Amazon Polly to automate the creation of accessible audio descriptions for video content. This approach can significantly reduce the time and cost required to make videos accessible for visually impaired audiences.
- Training Llama 3.3 Swallow: A Japanese sovereign LLM on Amazon SageMaker HyperPodby Kazuki Fujii on June 13, 2025 at 3:54 pm
The Institute of Science Tokyo has successfully trained Llama 3.3 Swallow, a 70-billion-parameter large language model (LLM) with enhanced Japanese capabilities, using Amazon SageMaker HyperPod. The model demonstrates superior performance in Japanese language tasks, outperforming GPT-4o-mini and other leading models. This technical report details the training infrastructure, optimizations, and best practices developed during the project.
- Accelerating Articul8’s domain-specific model development with Amazon SageMaker HyperPodby Yashesh Shroff on June 12, 2025 at 5:48 pm
Learn how Articul8 is redefining enterprise generative AI with domain-specific models that outperform general-purpose LLMs in real-world applications. In our latest blog post, we dive into how Amazon SageMaker HyperPod accelerated the development of Articul8’s industry-leading semiconductor model—achieving 2X higher accuracy that top open source models while slashing deployment time by 4X.
- How VideoAmp uses Amazon Bedrock to power their media analytics interfaceby Suzanne Willard, Makoto Uchida on June 12, 2025 at 5:46 pm
In this post, we illustrate how VideoAmp, a media measurement company, worked with the AWS Generative AI Innovation Center (GenAIIC) team to develop a prototype of the VideoAmp Natural Language (NL) Analytics Chatbot to uncover meaningful insights at scale within media analytics data using Amazon Bedrock.
- Adobe enhances developer productivity using Amazon Bedrock Knowledge Basesby Kamran Razi on June 11, 2025 at 4:44 pm
Adobe partnered with the AWS Generative AI Innovation Center, using Amazon Bedrock Knowledge Bases and the Vector Engine for Amazon OpenSearch Serverless. This solution dramatically improved their developer support system, resulting in a 20% increase in retrieval accuracy. In this post, we discuss the details of this solution and how Adobe enhances their developer productivity.
- Amazon Nova Lite enables Bito to offer a free tier option for its AI-powered code reviewsby Eshan Bhatnagar on June 11, 2025 at 4:33 pm
Bito is an innovative startup that creates AI agents for a broad range of software developers. In this post, we share how Bito is able to offer a free tier option for its AI-powered code reviews using Amazon Nova.
- How Gardenia Technologies helps customers create ESG disclosure reports 75% faster using agentic generative AI on Amazon Bedrockby Federico Thibaud, Neil Holloway, Fraser Price, Christian Dunn and Frederica Schrager on June 11, 2025 at 4:23 pm
Gardenia Technologies, a data analytics company, partnered with the AWS Prototyping and Cloud Engineering (PACE) team to develop Report GenAI, a fully automated ESG reporting solution powered by the latest generative AI models on Amazon Bedrock. This post dives deep into the technology behind an agentic search solution using tooling with Retrieval Augmented Generation (RAG) and text-to-SQL capabilities to help customers reduce ESG reporting time by up to 75%. We demonstrate how AWS serverless technology, combined with agents in Amazon Bedrock, are used to build scalable and highly flexible agent-based document assistant applications.
- NVIDIA Nemotron Super 49B and Nano 8B reasoning models now available in Amazon Bedrock Marketplace and Amazon SageMaker JumpStartby Niithiyn Vijeaswaran on June 11, 2025 at 12:00 pm
The Llama 3.3 Nemotron Super 49B V1 and Llama 3.1 Nemotron Nano 8B V1 are now available in Amazon Bedrock Marketplace and Amazon SageMaker JumpStart. With this launch, you can now deploy NVIDIA’s newest reasoning models to build, experiment, and responsibly scale your generative AI ideas on AWS.
- Automate customer support with Amazon Bedrock, LangGraph, and Mistral modelsby Deepesh Dhapola on June 10, 2025 at 4:19 pm
In this post, we demonstrate how to use Amazon Bedrock and LangGraph to build a personalized customer support experience for an ecommerce retailer. By integrating the Mistral Large 2 and Pixtral Large models, we guide you through automating key customer support workflows such as ticket categorization, order details extraction, damage assessment, and generating contextual responses.
- Build responsible AI applications with Amazon Bedrock Guardrailsby Divya Muralidharan on June 10, 2025 at 4:16 pm
In this post, we demonstrate how Amazon Bedrock Guardrails helps block harmful and undesirable multimodal content. Using a healthcare insurance call center scenario, we walk through the process of configuring and testing various guardrails.
- Effective cost optimization strategies for Amazon Bedrockby Biswanath Mukherjee on June 10, 2025 at 4:13 pm
With the increasing adoption of Amazon Bedrock, optimizing costs is a must to help keep the expenses associated with deploying and running generative AI applications manageable and aligned with your organization’s budget. In this post, you’ll learn about strategic cost optimization techniques while using Amazon Bedrock.
- How E.ON plans to save £10 million annually with AI diagnostics for smart meters powered by Amazon Textractby Sam Charlton on June 10, 2025 at 4:11 pm
E.ON’s story highlights how a creative application of Amazon Textract, combined with custom image analysis and pulse counting, can solve a real-world challenge at scale. By diagnosing smart meter errors through brief smartphone videos, E.ON aims to lower costs, improve customer satisfaction, and enhance overall energy service reliability. In this post, we dive into how this solution works and the impact it’s making.
- Building intelligent AI voice agents with Pipecat and Amazon Bedrock – Part 1by Adithya Suresh on June 9, 2025 at 3:50 pm
In this series of posts, you will learn how to build intelligent AI voice agents using Pipecat, an open-source framework for voice and multimodal conversational AI agents, with foundation models on Amazon Bedrock. It includes high-level reference architectures, best practices and code samples to guide your implementation.
- Stream multi-channel audio to Amazon Transcribe using the Web Audio APIby Jorge Lanzarotti on June 9, 2025 at 3:47 pm
In this post, we explore the implementation details of a web application that uses the browser’s Web Audio API and Amazon Transcribe streaming to enable real-time dual-channel transcription. By using the combination of AudioContext, ChannelMergerNode, and AudioWorklet, we were able to seamlessly process and encode the audio data from two microphones before sending it to Amazon Transcribe for transcription.
- How Kepler democratized AI access and enhanced client services with Amazon Q Businessby Evan Miller, Noah Kershaw, and Valerie Renda on June 9, 2025 at 3:39 pm
At Kepler, a global full-service digital marketing agency serving Fortune 500 brands, we understand the delicate balance between creative marketing strategies and data-driven precision. In this post, we share how implementing Amazon Q Business transformed our operations by democratizing AI access across our organization while maintaining stringent security standards, resulting in an average savings of 2.7 hours per week per employee in manual work and improved client service delivery.
- Build a serverless audio summarization solution with Amazon Bedrock and Whisperby Kaiyin Hu on June 6, 2025 at 5:34 pm
In this post, we demonstrate how to use the Open AI Whisper foundation model (FM) Whisper Large V3 Turbo, available in Amazon Bedrock Marketplace, which offers access to over 140 models through a dedicated offering, to produce near real-time transcription. These transcriptions are then processed by Amazon Bedrock for summarization and redaction of sensitive information.
- Implement semantic video search using open source large vision models on Amazon SageMaker and Amazon OpenSearch Serverlessby Alexander Arzhanov on June 6, 2025 at 5:13 pm
In this post, we demonstrate how to use large vision models (LVMs) for semantic video search using natural language and image queries. We introduce some use case-specific methods, such as temporal frame smoothing and clustering, to enhance the video search performance. Furthermore, we demonstrate the end-to-end functionality of this approach by using both asynchronous and real-time hosting options on Amazon SageMaker AI to perform video, image, and text processing using publicly available LVMs on the Hugging Face Model Hub. Finally, we use Amazon OpenSearch Serverless with its vector engine for low-latency semantic video search.