Key Moments
AI Dev 25 | Mike Chambers: Serverless Agentic Workflows with Amazon Bedrock
Key Moments
Build scalable, secure serverless AI agents with AWS Bedrock, integrating tools, code interpreters, and knowledge bases for production.
Key Insights
Amazon Bedrock provides a managed service for building and deploying AI agents, abstracting infrastructure complexity.
Agentic workflows leverage LLMs and tools (like Lambda functions) to interact with external systems and perform tasks.
AWS services such as Lambda, S3, DynamoDB, and OpenSearch are integral to building robust and scalable agentic systems.
Key features for production-ready agents include managed tools, a code interpreter for deterministic tasks, guardrails for safety, and knowledge bases for data access.
Newer features like inline agents and Amazon's Nova models offer faster development cycles and enhanced capabilities.
Serverless architecture is crucial for scalability and managing infrastructure without manual intervention.
INTRODUCTION TO AGENTIC WORKFLOWS AND PRODUCTION DEPLOYMENT
The session introduces agentic workflows, defined as systems that use generative AI to connect with external systems. The primary focus is on moving these agents from local development to production environments, emphasizing the importance of scalability, security, and manageable infrastructure. This transition is a growing concern for organizations looking to realize tangible outcomes from their AI investments.
AMAZON BEDROCK AND AWS SERVICES FOR AGENTS
Amazon Bedrock serves as a fully managed service for building generative AI applications, offering access to a variety of foundation models from different providers, including Amazon's own Nova models, Anthropic, and Meta. It allows seamless integration with other AWS services like Lambda for running code snippets at scale, S3 and DynamoDB for data storage, and managed components for agent configuration, retrieval augmented generation (RAG), and guardrails.
CORE COMPONENTS OF AN AGENTIC WORKFLOW
An agent acts as an orchestrator powered by a large language model, utilizing tools (represented as external systems) to interact with the outside world. Key to agentic workflows is the ability to self-loop, allowing the agent to chain tool executions and reasoning steps. For production, these agents can be configured and managed through Bedrock, ensuring scalability and reliability without manual infrastructure management.
ENABLING AGENTS WITH TOOLS AND FUNCTIONALITY
To extend agent capabilities, developers can define 'action groups' that map to external services, such as AWS Lambda functions. Each action group contains specific actions (tools) with detailed schemas describing their purpose and parameters. This allows the LLM, through the agent, to understand and invoke these tools, enabling it to perform complex queries, retrieve data, or trigger backend processes like customer support interactions.
ADVANCED CAPABILITIES: CODE INTERPRETER AND GUARDRAILS
For tasks requiring deterministic execution, such as mathematical calculations or data manipulation, Bedrock agents can be equipped with a code interpreter. This secure, managed environment allows the agent to write and execute code, returning results to the LLM. Additionally, guardrails are implemented to monitor input and output, ensuring policy compliance, preventing prompt attacks, and maintaining model safety and reliability, crucial for customer-facing applications.
KNOWLEDGE BASES AND RETRIEVAL AUGMENTED GENERATION (RAG)
Fully managed knowledge bases enable agents to access proprietary data. Documents are stored in S3, then processed by Bedrock Knowledge Base to be chunked, embedded into vectors, and stored in a vector store like Amazon OpenSearch Serverless. This integrates retrieval augmented generation (RAG) capabilities, allowing agents to provide informed responses based on up-to-date company information, greatly enhancing their utility for tasks like customer support.
INNOVATIONS: AMAZON NOVA MODELS AND INLINE AGENTS
Recent advancements include Amazon's own Nova models, offering improved price-performance for generative AI tasks. Furthermore, the introduction of 'inline agents' simplifies development by allowing agent configurations to be invoked directly within a development environment without full cloud deployment. This approach accelerates iteration cycles and opens possibilities for dynamic agent configuration generation and local tool execution.
PRODUCTION READINESS AND SERVERLESS ARCHITECTURE
The entire architecture discussed emphasizes serverless components, meaning there is no infrastructure to manage, patch, or scale manually. Services like Lambda, Bedrock Agents, and managed databases automatically scale with demand. This serverless approach is fundamental to achieving production-ready AI agents that can reliably handle varying workloads and deliver consistent performance.
Mentioned in This Episode
●Products
●Software & Apps
●Organizations
●Books
●Concepts
Key Steps for Deploying Serverless Agentic Workflows
Practical takeaways from this episode
Do This
Avoid This
Common Questions
An agent is a system that uses generative AI to connect with external systems and tools beyond its internal model capabilities. It acts as an orchestrator that can loop back to itself, interact with tools, and reason through a process to achieve a goal.
Topics
Mentioned in this video
LLMs from Anthropic, available via Amazon Bedrock.
A managed AWS service used as the vector store for the knowledge base.
A specific model within the Amazon Nova family, used as an example for inline agents.
The part of Bedrock used to invoke and interact with agents.
A comprehensive service for building generative AI applications, offering various foundation models and managed components.
The specific Amazon Bedrock service for creating and managing agents.
Proprietary LLMs developed by Amazon, available through Bedrock.
Amazon's flagship family of LLMs, mentioned as an improvement and alternative to previous models.
The AWS SDK for Python, used to interact with AWS services from Python code.
A use case for the agent being built, focused on handling customer inquiries about mugs.
Components that agents use to interact with external systems.
A feature allowing agents to be invoked directly from configuration without full cloud deployment, enabling faster development.
More from DeepLearningAI
View all 65 summaries
1 minThe #1 Skill Employers Want in 2026
1 minThe truth about tech layoffs and AI..
2 minBuild and Train an LLM with JAX
1 minWhat should you learn next? #AI #deeplearning
Found this useful? Build your knowledge library
Get AI-powered summaries of any YouTube video, podcast, or article in seconds. Save them to your personal pods and access them anytime.
Try Summify free