Solution of the Stochastic control problem in unbounded domains
โ Scribed by Prentiss Robinson; John Moore
- Publisher
- Elsevier Science
- Year
- 1973
- Tongue
- English
- Weight
- 541 KB
- Volume
- 295
- Category
- Article
- ISSN
- 0016-0032
No coin nor oath required. For personal study only.
โฆ Synopsis
Bellman's dynamic programming equation for the optimal index and control law for stochastic control problems is a parabolic or elliptic partial differential equation frequently defined in an unbounded domain. Existing methods of solution require bounded domain approximations, the application of singular perturbation techniques or Monte Carlo simulation procedures. In this paper, using the fact that Poisson impulse noise tends to a Gaussian process under certain limiting conditions, a method which achieves an arbitrarily good approximate solution to the &ocha&ic control problem is given. The method uses the two iterative techniques of successive approximation and quasi-linearization and is inherently more eficient than existing methods of solution.
๐ SIMILAR VOLUMES
TABLE 3 Simulation Results of Parallel Adding Method Number of P -Code 1 2 3 3 Successful Number of P -Code 15044 77 0 7 Total Number of P -Code 823543 823543 823543 7 creases. The interference between the P -code and P -code 7 3 is more serious in VAD. That is, the interference from K is more impo