Hands-On LLM Serving and Optimization: Hosting Llms at Scale
暫譯: 實戰 LLM 服務與優化:大規模托管 LLMs

Wang, Chi, Hu, Peiheng

  • 出版商: O'Reilly
  • 出版日期: 2026-06-02
  • 售價: $2,700
  • 貴賓價: 9.8$2,646
  • 語言: 英文
  • 頁數: 371
  • 裝訂: Quality Paper - also called trade paper
  • ISBN: 9798341621497
  • ISBN-13: 9798341621497
  • 相關分類: Large language model
  • 海外代購書籍(需單獨結帳)

相關主題

商品描述

Large language models (LLMs) are rapidly becoming the backbone of AI-driven applications. Without proper optimization, however, LLMs can be expensive to run, slow to serve, and prone to performance bottlenecks. As the demand for real-time AI applications grows, along comes Hands-On Serving and Optimizing LLM Models, a comprehensive guide to the complexities of deploying and optimizing LLMs at scale.

In this hands-on book, authors Chi Wang and Peiheng Hu take a real-world approach backed by practical examples and code, and assemble essential strategies for designing robust infrastructures that are equal to the demands of modern AI applications. Whether you're building high-performance AI systems or looking to enhance your knowledge of LLM optimization, this indispensable book will serve as a pillar of your success.

  • Learn the key principles for designing a model-serving system tailored to popular business scenarios
  • Understand the common challenges of hosting LLMs at scale while minimizing costs
  • Pick up practical techniques for optimizing LLM serving performance
  • Build a model-serving system that meets specific business requirements
  • Improve LLM serving throughput and reduce latency
  • Host LLMs in a cost-effective manner, balancing performance and resource efficiency

商品描述(中文翻譯)

大型語言模型(LLMs)正迅速成為人工智慧驅動應用的基石。然而,若未經適當優化,LLMs 的運行成本可能高昂、服務速度緩慢,且容易出現性能瓶頸。隨著對即時人工智慧應用需求的增長,《Hands-On Serving and Optimizing LLM Models》一書應運而生,這是一本全面指南,探討在大規模部署和優化 LLMs 的複雜性。

在這本實作導向的書中,作者 Chi Wang 和 Peiheng Hu 採取了基於實際案例和程式碼的現實世界方法,並組合出設計堅固基礎設施的基本策略,以滿足現代人工智慧應用的需求。無論您是構建高性能的人工智慧系統,還是希望增強對 LLM 優化的知識,這本不可或缺的書籍將成為您成功的支柱。

- 學習針對流行商業場景設計模型服務系統的關鍵原則
- 理解在大規模托管 LLMs 時常見的挑戰,同時最小化成本
- 獲取優化 LLM 服務性能的實用技術
- 建立滿足特定商業需求的模型服務系統
- 改善 LLM 服務的吞吐量並降低延遲
- 以具成本效益的方式托管 LLMs,平衡性能和資源效率