Dynamic programming for ergodic control with partial observations
A dynamic programming principle is derived for a discrete time Markov control process taking values in a finite dimensional space, with ergodic cost and partial observations. This uses the embedding of the process into another for which an accessible atom exists and hence a coupling argument can be used. In turn, this is used for deriving a martingale dynamic programming principle for ergodic control of partially observed diffusion processes, by 'lifting' appropriate estimates from a discrete time problem associated with it to the continuous time problem.
Year of publication: |
2003
|
---|---|
Authors: | Borkar, V. S. |
Published in: |
Stochastic Processes and their Applications. - Elsevier, ISSN 0304-4149. - Vol. 103.2003, 2, p. 293-310
|
Publisher: |
Elsevier |
Saved in:
Online Resource
Saved in favorites
Similar items by person
-
REINFORCEMENT LEARNING IN MARKOVIAN EVOLUTIONARY GAMES
BORKAR, V. S., (2002)
-
On De Finetti coherence and Kolmogorov probability
Borkar, V. S., (2004)
-
Bayesian learning and asymptotic equilibria in stochastic dynamic games
Borkar, Vivek S., (1994)
- More ...