Approximate Dynamic Programming: Solving the Curses of Dimensionality / Edition 1

Hardcover (Print)
Buy New
Buy New from
Used and New from Other Sellers
Used and New from Other Sellers
from $175.00
Usually ships in 1-2 business days
Other sellers (Hardcover)
  • All (2) from $175.00   
  • New (1) from $175.00   
  • Used (1) from $0.00   


A complete and accessible introduction to the real-world applications of approximate dynamic programming

With the growing levels of sophistication in modern-day operations, it is vital for practitioners to understand how to approach, model, and solve complex industrial problems. Approximate Dynamic Programming is a result of the author's decades of experience working in large industrial settings to develop practical and high-quality solutions to problems that involve making decisions in the presence of uncertainty. This groundbreaking book uniquely integrates four distinct disciplines—Markov design processes, mathematical programming, simulation, and statistics—to demonstrate how to successfully model and solve a wide range of real-life problems using the techniques of approximate dynamic programming (ADP). The reader is introduced to the three curses of dimensionality that impact complex problems and is also shown how the post-decision state variable allows for the use of classical algorithmic strategies from operations research to treat complex stochastic optimization problems.

Designed as an introduction and assuming no prior training in dynamic programming of any form, Approximate Dynamic Programming contains dozens of algorithms that are intended to serve as a starting point in the design of practical solutions for real problems. The book provides detailed coverage of implementation challenges including: modeling complex sequential decision processes under uncertainty, identifying robust policies, designing and estimating value function approximations, choosing effective stepsize rules, and resolving convergence issues.

With a focus on modeling and algorithms in conjunction with the language of mainstream operations research, artificial intelligence, and control theory, Approximate Dynamic Programming:

  • Models complex, high-dimensional problems in a natural and practical way, which draws on years of industrial projects
  • Introduces and emphasizes the power of estimating a value function around the post-decision state, allowing solution algorithms to be broken down into three fundamental steps: classical simulation, classical optimization, and classical statistics
  • Presents a thorough discussion of recursive estimation, including fundamental theory and a number of issues that arise in the development of practical algorithms
  • Offers a variety of methods for approximating dynamic programs that have appeared in previous literature, but that have never been presented in the coherent format of a book

Motivated by examples from modern-day operations research, Approximate Dynamic Programming is an accessible introduction to dynamic modeling and is also a valuable guide for the development of high-quality solutions to problems that exist in operations research and engineering. The clear and precise presentation of the material makes this an appropriate text for advanced undergraduate and beginning graduate courses, while also serving as a reference for researchers and practitioners. A companion Web site is available for readers, which includes additional exercises, solutions to exercises, and data sets to reinforce the book's main concepts.

Read More Show Less

Editorial Reviews

From the Publisher
"Perhaps the most appealing aspect of Professor Powell’s book is the fact that it spans both theory and practice...Problems, deemed intractable a few years ago, are now easily solved by using the exhibited techniques in this book. I would strongly recommend the book to any practitioner facing complex, dynamic models involving constantly changing information streams." (IIE Transactions-Operations Engineering, 2008) 

"Focus[es] on the core … of dynamic programming with a simple and clear exposition of the material … while … elevating the standard of the theory."*(Computing Reviews, May 5, 2008)

 "Motivated by examples from modern-day operations research, Approximate Dynamic Programming is an accessible introduction to dynamic modeling and is also a valuable guide for the development of high-quality solutions to problems that exist in operations research and engineering.  (Mathematical Reviews, 2008)

Read More Show Less

Product Details

  • ISBN-13: 9780470171554
  • Publisher: Wiley, John & Sons, Incorporated
  • Publication date: 10/5/2007
  • Series: Wiley Series in Probability and Statistics Series , #703
  • Edition description: Older Edition
  • Edition number: 1
  • Pages: 488
  • Product dimensions: 7.24 (w) x 10.20 (h) x 1.24 (d)

Meet the Author

Warren B. Powell, PhD, is Professor of Operations Research and Financial Engineering at Princeton University, where he is founder and Director of CASTLE Laboratory, a research unit that works with industrial partners to test new ideas found in operations research. The recipient of the 2004 INFORMS Fellow Award, Dr. Powell has authored over 100 refereed publications on stochastic optimization, approximate dynamic programming, and dynamic resource management.

Read More Show Less

Table of Contents



1. The challenges of dynamic programming.

1.1 A dynamic programming example: a shortest path problem.

1.2 The three curses of dimensionality.

1.3 Some real applications.

1.4 Problem classes.

1.5 The many dialects of dynamic programming.

1.6 What is new in this book?

1.7 Bibliographic notes.

2. Some illustrative models.

2.1 Deterministic problems.

2.2 Stochastic problems.

2.3 Information acquisition problems.

2.4 A simple modeling framework for dynamic programs.

2.5 Bibliographic notes.


3. Introduction to Markov decision processes.

3.1 The optimality equations.

3.2 Finite horizon problems.

3.3 Infinite horizon problems.

3.4 Value iteration.

3.5 Policy iteration.

3.6 Hybrid valuepolicy iteration.

3.7 The linear programming method for dynamic programs.

3.8 Monotone policies.

3.9 Why does it work?

3.10 Bibliographic notes.


4. Introduction to approximate dynamic programming.

4.1 The three curses of dimensionality (revisited).

4.2 The basic idea.

4.3 Sampling random variables .

4.4 ADP using the postdecision state variable.

4.5 Lowdimensional representations of value functions.

4.6 So just what is approximate dynamic programming?

4.7 Experimental issues.

4.8 Dynamic programming with missing or incomplete models.

4.9 Relationship to reinforcement learning.

4.10 But does it work?

4.11 Bibliographic notes.


5. Modeling dynamic programs.

5.1 Notational style.

5.2 Modeling time.

5.3 Modeling resources.

5.4 The states of our system.

5.5 Modeling decisions.

5.6 The exogenous information process.

5.7 The transition function.

5.8 The contribution function.

5.9 The objective function.

5.10 A measuretheoretic view of information.

5.11 Bibliographic notes.


6. Stochastic approximation methods.

6.1 A stochastic gradient algorithm.

6.2 Some stepsize recipes.

6.3 Stochastic stepsizes.

6.4 Computing bias and variance.

6.5 Optimal stepsizes.

6.6 Some experimental comparisons of stepsize formulas.

6.7 Convergence.

6.8 Why does it work?

6.9 Bibliographic notes.


7. Approximating value functions.

7.1 Approximation using aggregation.

7.2 Approximation methods using regression models.

7.3 Recursive methods for regression models.

7.4 Neural networks.

7.5 Batch processes.

7.6 Why does it work?

7.7 Bibliographic notes.


8. ADP for finite horizon problems.

8.1 Strategies for finite horizon problems.

8.2 Qlearning.

8.3 Temporal difference learning.

8.4 Policy iteration.

8.5 Monte Carlo value and policy iteration.

8.6 The actorcritic paradigm.

8.7 Bias in value function estimation.

8.8 State sampling strategies.

8.9 Starting and stopping.

8.10 A taxonomy of approximate dynamic programming strategies.

8.11 Why does it work?

8.12 Bibliographic notes.


9. Infinite horizon problems.

9.1 From finite to infinite horizon.

9.2 Algorithmic strategies.

9.3 Stepsizes for infinite horizon problems.

9.4 Error measures.

9.5 Direct ADP for online applications.

9.6 Finite horizon models for steady state applications.

9.7 Why does it work?

9.8 Bibliographic notes.


10. Exploration vs. exploitation.

10.1 A learning exercise: the nomadic trucker.

10.2 Learning strategies.

10.3 A simple information acquisition problem.

10.4 Gittins indices and the information acquisition problem.

10.5 Variations.

10.6 The knowledge gradient algorithm.

10.7 Information acquisition in dynamic programming.

10.8 Bibliographic notes.


11. Value function approximations for special functions.

11.1 Value functions versus gradients.

11.2 Linear approximations.

11.3 Piecewise linear approximations.

11.4 The SHAPE algorithm.

11.5 Regression methods.

11.6 Cutting planes.

11.7 Why does it work?

11.8 Bibliographic notes.


12. Dynamic resource allocation.

12.1 An asset acquisition problem.

12.2 The blood management problem.

12.3 A portfolio optimization problem.

12.4 A general resource allocation problem.

12.5 A fleet management problem.

12.6 A driver management problem.

12.7 Bibliographic references.


13. Implementation challenges.

13.1 Will ADP work for your problem?

13.2 Designing an ADP algorithm for complex problems.

13.3 Debugging an ADP algorithm.

13.4 Convergence issues.

13.5 Modeling your problem.

13.6 Online vs. offline models.

13.7 If it works, patent it!

Read More Show Less

Customer Reviews

Be the first to write a review
( 0 )
Rating Distribution

5 Star


4 Star


3 Star


2 Star


1 Star


Your Rating:

Your Name: Create a Pen Name or

Barnes & Review Rules

Our reader reviews allow you to share your comments on titles you liked, or didn't, with others. By submitting an online review, you are representing to Barnes & that all information contained in your review is original and accurate in all respects, and that the submission of such content by you and the posting of such content by Barnes & does not and will not violate the rights of any third party. Please follow the rules below to help ensure that your review can be posted.

Reviews by Our Customers Under the Age of 13

We highly value and respect everyone's opinion concerning the titles we offer. However, we cannot allow persons under the age of 13 to have accounts at or to post customer reviews. Please see our Terms of Use for more details.

What to exclude from your review:

Please do not write about reviews, commentary, or information posted on the product page. If you see any errors in the information on the product page, please send us an email.

Reviews should not contain any of the following:

  • - HTML tags, profanity, obscenities, vulgarities, or comments that defame anyone
  • - Time-sensitive information such as tour dates, signings, lectures, etc.
  • - Single-word reviews. Other people will read your review to discover why you liked or didn't like the title. Be descriptive.
  • - Comments focusing on the author or that may ruin the ending for others
  • - Phone numbers, addresses, URLs
  • - Pricing and availability information or alternative ordering information
  • - Advertisements or commercial solicitation


  • - By submitting a review, you grant to Barnes & and its sublicensees the royalty-free, perpetual, irrevocable right and license to use the review in accordance with the Barnes & Terms of Use.
  • - Barnes & reserves the right not to post any review -- particularly those that do not follow the terms and conditions of these Rules. Barnes & also reserves the right to remove any review at any time without notice.
  • - See Terms of Use for other conditions and disclaimers.
Search for Products You'd Like to Recommend

Recommend other products that relate to your review. Just search for them below and share!

Create a Pen Name

Your Pen Name is your unique identity on It will appear on the reviews you write and other website activities. Your Pen Name cannot be edited, changed or deleted once submitted.

Your Pen Name can be any combination of alphanumeric characters (plus - and _), and must be at least two characters long.

Continue Anonymously
Sort by: Showing 1 Customer Reviews
  • Anonymous

    Posted February 9, 2009

    No text was provided for this review.

Sort by: Showing 1 Customer Reviews

If you find inappropriate content, please report it to Barnes & Noble
Why is this product inappropriate?
Comments (optional)