Hands-On LLM Serving and Optimization: Hosting Llms at Scale
暫譯: 實戰 LLM 服務與優化:大規模托管 LLMs
Wang, Chi, Hu, Peiheng
- 出版商: O'Reilly
- 出版日期: 2026-06-02
- 售價: $2,700
- 貴賓價: 9.8 折 $2,646
- 語言: 英文
- 頁數: 371
- 裝訂: Quality Paper - also called trade paper
- ISBN: 9798341621497
- ISBN-13: 9798341621497
-
相關分類:
Large language model
海外代購書籍(需單獨結帳)
相關主題
商品描述
Large language models (LLMs) are rapidly becoming the backbone of AI-driven applications. Without proper optimization, however, LLMs can be expensive to run, slow to serve, and prone to performance bottlenecks. As the demand for real-time AI applications grows, along comes Hands-On Serving and Optimizing LLM Models, a comprehensive guide to the complexities of deploying and optimizing LLMs at scale.
In this hands-on book, authors Chi Wang and Peiheng Hu take a real-world approach backed by practical examples and code, and assemble essential strategies for designing robust infrastructures that are equal to the demands of modern AI applications. Whether you're building high-performance AI systems or looking to enhance your knowledge of LLM optimization, this indispensable book will serve as a pillar of your success.
- Learn the key principles for designing a model-serving system tailored to popular business scenarios
- Understand the common challenges of hosting LLMs at scale while minimizing costs
- Pick up practical techniques for optimizing LLM serving performance
- Build a model-serving system that meets specific business requirements
- Improve LLM serving throughput and reduce latency
- Host LLMs in a cost-effective manner, balancing performance and resource efficiency
商品描述(中文翻譯)
大型語言模型(LLMs)正迅速成為人工智慧驅動應用的基石。然而,若未經適當優化,LLMs 的運行成本可能高昂、服務速度緩慢,且容易出現性能瓶頸。隨著對即時人工智慧應用需求的增長,《Hands-On Serving and Optimizing LLM Models》一書應運而生,這是一本全面指南,探討在大規模部署和優化 LLMs 的複雜性。
在這本實作導向的書中,作者 Chi Wang 和 Peiheng Hu 採取了基於實際案例和程式碼的現實世界方法,並組合出設計堅固基礎設施的基本策略,以滿足現代人工智慧應用的需求。無論您是構建高性能的人工智慧系統,還是希望增強對 LLM 優化的知識,這本不可或缺的書籍將成為您成功的支柱。
- 學習針對流行商業場景設計模型服務系統的關鍵原則
- 理解在大規模托管 LLMs 時常見的挑戰,同時最小化成本
- 獲取優化 LLM 服務性能的實用技術
- 建立滿足特定商業需求的模型服務系統
- 改善 LLM 服務的吞吐量並降低延遲
- 以具成本效益的方式托管 LLMs,平衡性能和資源效率