Agentic AI Architecture

This course is a hands-on, three-day course for experienced developers building production-ready, multi-agent AI systems. Participants learn to architect agent workflows with MCP tools, RAG grounding, orchestration, human-in-the-loop approvals, observability, evaluation, and security, culminating in a realistic enterprise capstone suitable for real-world deployment.

Description

This instructor-led course teaches experienced developers how to design, build, and deploy production-ready multi-agent AI systems. The program combines hands-on implementation with architectural best practices, focusing on tool integration, retrieval-augmented generation, orchestration patterns, and enterprise-grade concerns like observability, evaluation, and security.

Participants build a complete Event Operations Planner — a multi-agent system that coordinates specialized AI agents to plan safe, compliant events with human-in-the-loop approvals.

Learning Objectives

By completing this course, participants will be able to:

  • Architect multi-agent systems using the supervisor orchestration pattern
  • Engineer tool-calling agents using the Model Context Protocol (MCP)
  • Build RAG systems with semantic search for policy grounding
  • Orchestrate agent workflows with LangGraph state machines
  • Add structured observability and distributed tracing to agentic systems
  • Create human-in-the-loop approval workflows with revision loops
  • Evaluate agent behavior using functional, RAG, and LLM-as-judge evaluators
  • Secure agents against prompt injection and implement defense-in-depth

Course Modules

Day 1: Foundations

Lab 1: Tools & MCP — Building the foundation for agent-world interaction through FastAPI endpoints and the Model Context Protocol. Students implement venue and weather APIs, then wrap them as MCP tools that agents can discover and invoke.

Lab 2: Retrieval-Augmented Generation — Grounding agent decisions in enterprise knowledge. Students build a complete RAG pipeline: document chunking, FAISS vector indexing, semantic search, and retrieval formatting for LLM consumption.

Lab 3: Single Agent Anatomy — Understanding the core agent loop pattern. Students implement a Policy/RAG Agent that searches the knowledge base, extracts constraints with citations, and demonstrates the "retrieve or refuse" pattern for grounded behavior.

Day 2: Scale

Lab 4: Multi-Agent Orchestration — Coordinating multiple specialist agents using LangGraph. Students build a supervisor that routes between Intake, Policy, Planner, and Approvals agents based on a status-driven state machine.

Lab 5: Observability — Understanding agent behavior in production. Students add structured logging with correlation IDs, implement audit trails, and integrate distributed tracing for end-to-end visibility.

Lab 6: Human-in-the-Loop — Knowing when and how to involve humans. Students implement dual-approval workflows requiring both Security Lead and Operations Lead sign-off, with revision loops that route rejected plans back for improvement.

Day 3: Production

Lab 7: Evaluation — Testing agentic systems systematically. Students build an evaluation framework with functional evaluators (status checks, constraint validation), RAG evaluators (faithfulness, context precision), and LLM-as-judge evaluators for subjective quality assessment.

Lab 8: Security & Governance — Protecting agentic systems from adversarial inputs. Students implement layered prompt injection detection using pattern matching and risk scoring, configure block vs. log-only modes, and apply defense-in-depth principles.

Technology Stack

ComponentTechnologyPurpose
Language Python 3.11+ Core development
Multi-Agent LangGraph Workflow orchestration
APIs FastAPI Domain services
Tool Protocol MCP (Anthropic) Agent-tool communication
LLM LiteLLM Model-agnostic LLM calls
RAG FAISS + sentence-transformers Vector search
Logging structlog Structured observability
Tracing Arize Phoenix Distributed tracing

Capstone Project

The Event Operations Planner capstone brings all course concepts together in a realistic, production-grade agentic system. Participants build a multi-agent application that coordinates intake, policy discovery, risk assessment, planning, and approvals to produce safe, compliant event plans. The system integrates real tools via MCP, grounds decisions using retrieval-augmented generation, routes work through a supervisor-driven workflow, and enforces human-in-the-loop approvals from security and operations leaders. By the end of the course, learners have implemented a fully observable, evaluated, and secured agentic system that mirrors the challenges of deploying AI in real enterprise environments.

Audience

Python developers and ML engineers interested in building production-grade multi-agent AI systems. Ideal for teams planning to deploy agentic applications in enterprise environments where reliability, observability, and security are critical.

Materials Provided

  • Complete lab exercises with scaffolded code
  • Reference implementations (checkpoints) for each lab
  • Knowledge base documents (policies, templates, postmortems)
  • Evaluation scenarios and test suites
  • Full capstone solution with Streamlit UIs

Prerequisites

  • Developers comfortable scripting and building automation in Python (or similar languages)
  • Intermediate knowledge of concpets such as async/await, classes, type hints
  • Basic understanding of REST APIs
  • Familiarity with LLMs and prompt engineering concepts
  • Development environment with Python 3.11+, Poetry, and an LLM API key (OpenAI or Anthropic)

Similar courses

This course is an intermediate, hands-on course for developers who want to build reliable AI assistants and use LLMs effectively as collaborative development tools. Participants learn to design assistant architectures, control model behavior, integrate tools and enterprise knowledge, and evaluate quality—while also developing practical AI pair-programming skills used in team environments with tools like Claude Code.

More Information