Loading...
Please wait, while we are loading the content...
Similar Documents
Convergence of Sample Path Optimal Policies for Stochastic Dynamic Programming
| Content Provider | Semantic Scholar |
|---|---|
| Author | Fu, Michael C. Jin, Xing |
| Copyright Year | 2005 |
| Abstract | We consider the solution of stochastic dynamic programs using sample path estimates. Applying the theory of large deviations, we derive probability error bounds associated with the convergence of the estimated optimal policy to the true optimal policy, for finite horizon problems. These bounds decay at an exponential rate, in contrast with the usual canonical (inverse) square root rate associated with estimation of the value (cost-to-go) function itself. These results have practical implications for Monte Carlo simulation-based solution approaches to stochastic dynamic programming problems where it is impractical to extract the explicit transition probabilities of the underlying system model. |
| File Format | PDF HTM / HTML |
| Alternate Webpage(s) | http://drum.lib.umd.edu/bitstream/handle/1903/6548/TR_2005-84.pdf?isAllowed=y&sequence=1 |
| Alternate Webpage(s) | http://www.dtic.mil/dtic/tr/fulltext/u2/a438510.pdf |
| Language | English |
| Access Restriction | Open |
| Subject Keyword | Convergence (action) Dynamic programming Estimated Extreme value theory Markov chain Monte Carlo method Probability Simulation Stochastic programming exponential |
| Content Type | Text |
| Resource Type | Article |