AWS Machine Learning Blog Official Machine Learning Blog of Amazon Web Services
- Unify structured data in Amazon Aurora and unstructured data in Amazon S3 for insights using Amazon Qby Monjumi Sarma on November 20, 2024 at 4:15 pm
In today’s data-intensive business landscape, organizations face the challenge of extracting valuable insights from diverse data sources scattered across their infrastructure. Whether it’s structured data in databases or unstructured content in document repositories, enterprises often struggle to efficiently query and use this wealth of information. In this post, we explore how you can use Amazon
- Automate Q&A email responses with Amazon Bedrock Knowledge Basesby Darrin Weber on November 20, 2024 at 4:11 pm
In this post, we illustrate automating the responses to email inquiries by using Amazon Bedrock Knowledge Bases and Amazon Simple Email Service (Amazon SES), both fully managed services. By linking user queries to relevant company domain information, Amazon Bedrock Knowledge Bases offers personalized responses.
- Streamline RAG applications with intelligent metadata filtering using Amazon Bedrockby Mani Khanuja on November 20, 2024 at 4:08 pm
In this post, we explore an innovative approach that uses LLMs on Amazon Bedrock to intelligently extract metadata filters from natural language queries. By combining the capabilities of LLM function calling and Pydantic data models, you can dynamically extract metadata from user queries. This approach can also enhance the quality of retrieved information and responses generated by the RAG applications.
- Embedding secure generative AI in mission-critical public safety applicationsby Lawrence Zorio III on November 20, 2024 at 3:59 pm
This post shows how Mark43 uses Amazon Q Business to create a secure, generative AI-powered assistant that drives operational efficiency and improves community service. We explain how they embedded Amazon Q Business web experience in their web application with low code, so they could focus on creating a rich AI experience for their customers.
- How FP8 boosts LLM training by 18% on Amazon SageMaker P5 instancesby Romil Shah on November 20, 2024 at 3:54 pm
LLM training has seen remarkable advances in recent years, with organizations pushing the boundaries of what’s possible in terms of model size, performance, and efficiency. In this post, we explore how FP8 optimization can significantly speed up large model training on Amazon SageMaker P5 instances.
- Racing into the future: How AWS DeepRacer fueled my AI and ML journeyby Matt Camp on November 19, 2024 at 8:52 pm
In 2018, I sat in the audience at AWS re:Invent as Andy Jassy announced AWS DeepRacer—a fully autonomous 1/18th scale race car driven by reinforcement learning. At the time, I knew little about AI or machine learning (ML). As an engineer transitioning from legacy networks to cloud technologies, I had never considered myself a developer.
- Your guide to generative AI and ML at AWS re:Invent 2024by Mukund Birje on November 19, 2024 at 8:25 pm
In this attendee guide, we’re highlighting a few of our favorite sessions to give you a glimpse into what’s in store. To help you plan your agenda for this year’s re:Invent, here are some highlights of the generative AI and ML sessions. Visit the session catalog to learn about all our generative AI and ML sessions.
- Customize small language models on AWS with automotive terminologyby Bruno Pistone on November 19, 2024 at 6:13 pm
In this post, we guide you through the phases of customizing SLMs on AWS, with a specific focus on automotive terminology for diagnostics as a Q&A task. We begin with the data analysis phase and progress through the end-to-end process, covering fine-tuning, deployment, and evaluation. We compare a customized SLM with a general purpose LLM, using various metrics to assess vocabulary richness and overall accuracy.
- Automate emails for task management using Amazon Bedrock Agents, Amazon Bedrock Knowledge Bases, and Amazon Bedrock Guardrailsby Manu Mishra on November 19, 2024 at 6:05 pm
In this post, we demonstrate how to create an automated email response solution using Amazon Bedrock and its features, including Amazon Bedrock Agents, Amazon Bedrock Knowledge Bases, and Amazon Bedrock Guardrails.
- Automate building guardrails for Amazon Bedrock using test-driven developmentby Harsh Patel on November 19, 2024 at 5:57 pm
Amazon Bedrock Guardrails helps implement safeguards for generative AI applications based on specific use cases and responsible AI policies. Amazon Bedrock Guardrails assists in controlling the interaction between users and foundation models (FMs) by detecting and filtering out undesirable and potentially harmful content, while maintaining safety and privacy. In this post, we explore a solution that automates building guardrails using a test-driven development approach.
- Build cost-effective RAG applications with Binary Embeddings in Amazon Titan Text Embeddings V2, Amazon OpenSearch Serverless, and Amazon Bedrock Knowledge Basesby Shreyas Subramanian on November 18, 2024 at 8:09 pm
Today, we are happy to announce the availability of Binary Embeddings for Amazon Titan Text Embeddings V2 in Amazon Bedrock Knowledge Bases and Amazon OpenSearch Serverless. This post summarizes the benefits of this new binary vector support and gives you information on how you can get started.
- Automate cloud security vulnerability assessment and alerting using Amazon Bedrockby Shikhar Kwatra on November 18, 2024 at 8:08 pm
This post demonstrates a proactive approach for security vulnerability assessment of your accounts and workloads, using Amazon GuardDuty, Amazon Bedrock, and other AWS serverless technologies. This approach aims to identify potential vulnerabilities proactively and provide your users with timely alerts and recommendations, avoiding reactive escalations and other damages.
- DXC transforms data exploration for their oil and gas customers with LLM-powered toolsby Aude Genevay on November 18, 2024 at 8:05 pm
In this post, we show you how DXC and AWS collaborated to build an AI assistant using large language models (LLMs), enabling users to access and analyze different data types from a variety of data sources. The AI assistant is powered by an intelligent agent that routes user questions to specialized tools that are optimized for different data types such as text, tables, and domain-specific formats. It uses the LLM’s ability to understand natural language, write code, and reason about conversational context.
- How MSD uses Amazon Bedrock to translate natural language into SQL for complex healthcare databasesby Tesfagabir Meharizghi on November 18, 2024 at 6:57 pm
MSD, a leading pharmaceutical company, collaborates with AWS to implement a powerful text-to-SQL generative AI solution using Amazon Bedrock and Anthropic’s Claude 3.5 Sonnet model. This approach streamlines data extraction from complex healthcare databases like DE-SynPUF, enabling analysts to generate SQL queries from natural language questions. The solution addresses challenges such as coded columns, non-intuitive names, and ambiguous queries, significantly reducing query time and democratizing data access.
- Generate AWS Resilience Hub findings in natural language using Amazon Bedrockby Ibrahim Ahmad on November 18, 2024 at 6:51 pm
This blog post discusses a solution that combines AWS Resilience Hub and Amazon Bedrock to generate architectural findings in natural language. By using the capabilities of Resilience Hub and Amazon Bedrock, you can share findings with C-suite executives, engineers, managers, and other personas within your corporation to provide better visibility over maintaining a resilient architecture.
- Generate and evaluate images in Amazon Bedrock with Amazon Titan Image Generator G1 v2 and Anthropic Claude 3.5 Sonnetby Raul Tavares on November 18, 2024 at 6:33 pm
In this post, we demonstrate how to interact with the Amazon Titan Image Generator G1 v2 model on Amazon Bedrock to generate an image. Then, we show you how to use Anthropic’s Claude 3.5 Sonnet on Amazon Bedrock to describe it, evaluate it with a score from 1–10, explain the reason behind the given score, and suggest improvements to the image.
- How InsuranceDekho transformed insurance agent interactions using Amazon Bedrock and generative AIby Vishal Gupta on November 18, 2024 at 6:30 pm
In this post, we explain how InsuranceDekho harnessed the power of generative AI using Amazon Bedrock and Anthropic’s Claude to provide responses to customer queries on policy coverages, exclusions, and more. This let our customer care agents and POSPs confidently help our customers understand the policies without reaching out to insurance subject matter experts (SMEs) or memorizing complex plans while providing sales and after-sales services. The use of this solution has improved sales, cross-selling, and overall customer service experience.
- Considerations for addressing the core dimensions of responsible AI for Amazon Bedrock applicationsby Laura Verghote on November 15, 2024 at 7:18 pm
In this post, we introduce the core dimensions of responsible AI and explore considerations and strategies on how to address these dimensions for Amazon Bedrock applications.
- From RAG to fabric: Lessons learned from building real-world RAGs at GenAIIC – Part 2by Aude Genevay on November 15, 2024 at 5:22 pm
This post focuses on doing RAG on heterogeneous data formats. We first introduce routers, and how they can help managing diverse data sources. We then give tips on how to handle tabular data and will conclude with multimodal RAG, focusing specifically on solutions that handle both text and image data.
- Cohere Embed multimodal embeddings model is now available on Amazon SageMaker JumpStartby Breanne Warner on November 15, 2024 at 5:15 pm
The Cohere Embed multimodal embeddings model is now generally available on Amazon SageMaker JumpStart. This model is the newest Cohere Embed 3 model, which is now multimodal and capable of generating embeddings from both text and images, enabling enterprises to unlock real value from their vast amounts of data that exist in image form. In this post, we discuss the benefits and capabilities of this new model with some examples.