Loading...
Please wait, while we are loading the content...
Similar Documents
Interpolation-based Q-learning (2004)
| Content Provider | CiteSeerX |
|---|---|
| Author | Szepesvári, Csaba Smart, William D. |
| Description | Proceedings of the International Conference on Machine Learning |
| Abstract | We consider a variant of Q-learning in continuous state spaces under the total expected discounted cost criterion combined with local function approximation methods. Provided that the function approximator satisfies certain interpolation properties, the resulting algorithm is shown to converge with probability one. The limit function is shown to satisfy a fixed point equation of the Bellman type, where the fixed point operator depends on the stationary distribution of the exploration policy and the function approximation method. The basic algorithm is extended in several ways. In particular, a variant of the algorithm is obtained that is shown to converge in probability to the optimal Q function. Preliminary computer simulations are presented that confirm the validity of the approach. |
| File Format | |
| Publisher Date | 2004-01-01 |
| Access Restriction | Open |
| Subject Keyword | Exploration Policy Cost Criterion Fixed Point Equation Several Way Basic Algorithm Stationary Distribution Fixed Point Operator Optimal Function Function Approximation Method Preliminary Computer Simulation Limit Function Local Function Approximation Method Bellman Type Interpolation-based Q-learning Continuous State Space |
| Content Type | Text |
| Resource Type | Proceeding |