Dynamic programming and Markov processes
โ Scribed by Ronald A. Howard
- Publisher
- Technology Press of Massachusetts Institute of Technology
- Year
- 1960
- Tongue
- English
- Leaves
- 148
- Category
- Library
No coin nor oath required. For personal study only.
โฆ Table of Contents
Front cover
Dynamic Programming and Markov Processes
Copyright
Preface
Contents
Introduction
1 Markov Processes
The Toymaker Example-State Probabilities
The z- Transformation
z- Transform Analysis of Markov Processes
Transient, Multichain; and Periodic Behavior
2 Markov Processes with Rewards
Solution by Recurrence Relation
The Toymaker Example
z-Transform Analysis of the Markov Process with Rewards
Asymptotic Behavior
3 The Solution of the Sequential Decision Process by Value Iteration
Introduction of Alternatives
The Toymaker's Problem Solved by Value Iteration
Evaluation of the Value-Iteration Approach
4 The Policy-Iteration Method for the Solution of Sequential Decision Processes
The Value-Determination Operation
The Policy-Improvement Routine
The Iteration Cycle
The Toymaker's Problem
A Proof of the Properties of the Policy-Iteration Method
5 Use of the Policy-Iteration Method in Problems of Taxicab Operation, Baseball, and Automobile Replacement
An Example-Taxicab Operation
A Baseball Problem
The Policy-Improvement Routine
6 The Policy-Iteration Method for Multiple-Chain Processes
The Value-Determination Operation
The Policy-Improvement Routine
A Multichain Example
Properties of the Iteration Cycle
7 The Sequential Decision Process with Discounting
The Sequential Decision Process with Discounting Solved by Value Iteration
The Value-Determination Operation
The Policy-Improvement Routine
An Example
Proof of the Properties of the Iteration Cycle
The Sensitivity of the Optimal Policy to the Discount Factor
The Automobile Problem with Discounting
Summary
8 The Continuous- Time Decision Process
The Continuous-Time Markov Process
The Solution of Continuous-Time Markov Processes by Laplace Transformation
The Continuous-Time Markov Process with Rewards
The Continuous-Time Decision Problem
The Value-Determination Operation
The Policy-Improvement Routine
Completely Ergodic Processes
The Foreman's Dilemma
Computational Considerations
The Continuous-Time Decision Process with Discounting
Policy Improvement
An Example
Comparison with Discrete-Time Case
9 Conclusion
Appendix: The Relationship of Transient to Recurrent Behavior
References
General References
Index
Back cover
๐ SIMILAR VOLUMES
Markov processes --<br/> Markov processes with rewards --<br/> The solution of the sequential decision process by value iteration --<br/> The policy-iteration method for the solution of sequential --<br/> Use of the policy-iteration method in problems of taxicab operation, baseball, and automobi
The book presents an analytic structure for a decision-making system that is at the same time both general enough to be descriptive and yet computationally feasible. It is based on the Markov process as a system model, and uses and iterative technique like dynamic programming as its optimization met
The Wiley-Interscience Paperback Series consists of selected books that have been made more accessible to consumers in an effort to increase global appeal and general circulation. With these new unabridged softcover volumes, Wiley hopes to extend the lives of these works by making them available to
The Wiley-Interscience Paperback Series consists of selected books that have been made more accessible to consumers in an effort to increase global appeal and general circulation. With these new unabridged softcover volumes, Wiley hopes to extend the lives of these works by making them available to