Operational AI with Docker: Deploy, scale, and operate agentic AI services with Docker and Kubernetes
暫譯: 使用 Docker 的運營 AI:使用 Docker 和 Kubernetes 部署、擴展及運行自主 AI 服務
Raina, Ajeet Singh, Manvar, Harsh
- 出版商: Packt Publishing
- 出版日期: 2026-04-29
- 售價: $1,690
- 貴賓價: 9.5 折 $1,605
- 語言: 英文
- 頁數: 390
- 裝訂: Quality Paper - also called trade paper
- ISBN: 1807301095
- ISBN-13: 9781807301095
-
相關分類:
AI Coding、Docker
海外代購書籍(需單獨結帳)
相關主題
商品描述
Run production-grade GenAI workloads by containerizing, serving, and scaling LLMs, agents, and multi-model pipelines with Docker, MCP, and Kubernetes for cloud platforms
Key Features:
- Deploy and operate local and edge-friendly LLM inference using Docker Model Runner and an OpenAI-compatible API
- Orchestrate multi-model and multi-agent workloads with Docker Compose and Kubernetes patterns used by platform teams
- Purchase of the print or Kindle book includes a free PDF eBook
Book Description:
Modern AI systems don't fail at modeling; they fail in production. Moving from experiments to reliable, scalable systems requires more than notebooks and scripts. It requires infrastructure.
Operational AI with Docker shows you how to build, deploy, and operate AI systems that work beyond a single machine. You'll learn how to use Docker as a consistent runtime for machine learning workflows, package models as reproducible artifacts, and run them reliably across environments.
Starting with containerized machine learning, you'll progress to model serving, AI deployment, and scalable infrastructure using Kubernetes. You'll implement production-ready patterns for resource management, autoscaling, observability, and performance tuning, ensuring your AI workloads remain stable under real-world conditions.
The book goes beyond traditional MLOps by introducing agentic AI systems, including autonomous agents, multi-agent architectures, and secure execution environments. You'll also explore modern integration patterns using the Model Context Protocol (MCP), enabling AI systems to interact safely with tools, APIs, and data sources.
By the end of this book, you'll be able to design and operate production AI systems that are reproducible, scalable, and ready for real-world deployment using Docker and Kubernetes.
What You Will Learn:
- Containerize GenAI services using Docker images, registries, and Compose-based deployment stacks
- Package and distribute models as OCI artifacts for repeatable builds and controlled promotions across environments
- Choose GGUF quantization levels to balance cost, latency, and accuracy for cloud and hybrid runtimes
- Serve LLMs via Docker Model Runner with an OpenAI-compatible API suitable for internal platforms
- Integrate tools and data securely using MCP and Docker MCP Gateway with least-privilege access patterns
Who this book is for:
Cloud engineers, DevOps engineers, SREs, and platform engineers who need to deploy, operate, and scale GenAI workloads using Docker and Kubernetes on cloud, hybrid, or edge environments. You should be comfortable with the command line and basic service operations; prior Docker or Kubernetes exposure is helpful but not required.
Table of Contents
- Docker Desktop - The Runtime Foundation for AI/ML Workflows
- Understanding AI Models in Docker
- Model Service with Docker Model Runner
- Docker Offload for AI and ML Workflows
- Running ML Container Models on Kubernetes
- Protocol-Based AI Integration with MCP
- Building Autonomous AI Agents
- Multi-Model and Multi-Agent Architectures
- Advanced Agent Orchestration
商品描述(中文翻譯)
**運行生產級的 GenAI 工作負載,透過容器化、服務和擴展 LLM、代理和多模型管道,使用 Docker、MCP 和 Kubernetes 於雲端平台**
**主要特點:**
- 使用 Docker Model Runner 和與 OpenAI 兼容的 API 部署和操作本地及邊緣友好的 LLM 推論
- 使用平台團隊的 Docker Compose 和 Kubernetes 模式來協調多模型和多代理工作負載
- 購買印刷版或 Kindle 書籍包括免費 PDF 電子書
**書籍描述:**
現代 AI 系統在建模上不會失敗;它們在生產中會失敗。從實驗轉向可靠、可擴展的系統需要的不僅僅是筆記本和腳本。它需要基礎設施。
《使用 Docker 的運營 AI》將教你如何構建、部署和操作超越單一機器的 AI 系統。你將學會如何使用 Docker 作為機器學習工作流程的一致運行時,將模型打包為可重現的工件,並在不同環境中可靠地運行它們。
從容器化的機器學習開始,你將進展到模型服務、AI 部署和使用 Kubernetes 的可擴展基礎設施。你將實施生產就緒的資源管理、自動擴展、可觀察性和性能調整模式,確保你的 AI 工作負載在現實條件下保持穩定。
本書超越了傳統的 MLOps,介紹了代理 AI 系統,包括自主代理、多代理架構和安全執行環境。你還將探索使用模型上下文協議 (MCP) 的現代集成模式,使 AI 系統能夠安全地與工具、API 和數據源互動。
在本書結束時,你將能夠設計和操作可重現、可擴展並準備好在現實世界中部署的生產 AI 系統,使用 Docker 和 Kubernetes。
**你將學到的內容:**
- 使用 Docker 映像、註冊表和基於 Compose 的部署堆疊來容器化 GenAI 服務
- 將模型打包和分發為 OCI 工件,以便在不同環境中進行可重複的構建和控制的推廣
- 選擇 GGUF 量化級別,以平衡雲端和混合運行時的成本、延遲和準確性
- 通過 Docker Model Runner 以適合內部平台的與 OpenAI 兼容的 API 服務 LLM
- 使用 MCP 和 Docker MCP Gateway 以最小特權訪問模式安全地集成工具和數據
**本書適合誰:**
雲端工程師、DevOps 工程師、SRE 和平台工程師,他們需要在雲端、混合或邊緣環境中使用 Docker 和 Kubernetes 部署、操作和擴展 GenAI 工作負載。你應該對命令行和基本服務操作感到舒適;之前接觸 Docker 或 Kubernetes 是有幫助的,但不是必需的。
**目錄:**
- Docker Desktop - AI/ML 工作流程的運行時基礎
- 理解 Docker 中的 AI 模型
- 使用 Docker Model Runner 的模型服務
- AI 和 ML 工作流程的 Docker 卸載
- 在 Kubernetes 上運行 ML 容器模型
- 基於協議的 AI 集成與 MCP
- 構建自主 AI 代理
- 多模型和多代理架構
- 高級代理編排