Handbook of Learning and Approximate Dynamic Programming
Jennie Si, Andrew G. Barto, Warren Buckler Powell, Don Wunsch
- 出版商: Wiley
- 出版日期: 2004-08-02
- 定價: $6,600
- 售價: 9.5 折 $6,270
- 語言: 英文
- 頁數: 672
- 裝訂: Hardcover
- ISBN: 047166054X
- ISBN-13: 9780471660545
-
相關分類:
人工智慧、控制系統 Control-systems
立即出貨 (庫存=1)
買這商品的人也買了...
-
$1,176Optical Networks: A Practical Perspective, 2/e
-
$650$429 -
$2,030$1,929 -
$780$741 -
$931Network Systems Design Using Network Processors (Paperback)
-
$590$466 -
$490$387 -
$690$538 -
$990Beginning Visual C++ 6 (Paperback)
-
$550$468 -
$490$417 -
$4,740$4,503 -
$1,176Computer Organization and Design: The Hardware/Software Interface, 3/e(IE) (美國版ISBN:1558606041)
-
$760$646 -
$890$703 -
$650$553 -
$650$507 -
$1,615CCNA Cisco Certified Network Associate Study Guide, 5/e (640-801)
-
$880$748 -
$580$493 -
$2,375WiMAX Handbook
-
$420$328 -
$680$449 -
$450$356 -
$490$387
相關主題
商品描述
Description:
Approximate dynamic programming solves decision and control problems
While advances in science and engineering have enabled us to design and build complex systems, how to control and optimize them remains a challenge. This was made clear, for example, by the major power outage across dozens of cities in the Eastern United States and Canada in August of 2003. Learning and approximate dynamic programming (ADP) is emerging as one of the most promising mathematical and computational approaches to solve nonlinear, large-scale, dynamic control problems under uncertainty. It draws heavily both on rigorous mathematics and on biological inspiration and parallels, and helps unify new developments across many disciplines.
The foundations of learning and approximate dynamic programming have evolved from several fields–optimal control, artificial intelligence (reinforcement learning), operations research (dynamic programming), and stochastic approximation methods (neural networks). Applications of these methods span engineering, economics, business, and computer science. In this volume, leading experts in the field summarize the latest research in areas including:
- Reinforcement learning and its relationship to supervised learning
- Model-based adaptive critic designs
- Direct neural dynamic programming
- Hierarchical decision-making
- Multistage stochastic linear programming for resource allocation problems
- Concurrency, multiagency, and partial observability
- Backpropagation through time and derivative adaptive critics
- Applications of approximate dynamic programming and reinforcement learning in control-constrained agile missiles; power systems; heating, ventilation, and air conditioning; helicopter flight control; transportation and more.