COOKIES: By using this website you agree that we can place Google Analytics Cookies on your device for performance monitoring. |
University of Cambridge > Talks.cam > CUED Control Group Seminars > Q-learning and Pontryagin's Minimum Principle
Q-learning and Pontryagin's Minimum PrincipleAdd to your list(s) Download to your calendar using vCal
If you have a question about this talk, please contact Dr Ioannis Lestas. Q-learning and Pontryagin’s Minimum Principle Q-learning is a technique used to compute an optimal policy for a controlled Markov chain based on observations of the system controlled using a non-optimal policy. It has proven to be effective for models with finite state and action space. This paper establishes connections between Q-learning and nonlinear control of continuous-time models with general state space and general action space. The main contributions are summarized as follows.
This talk is part of the CUED Control Group Seminars series. This talk is included in these lists:
Note that ex-directory lists are not shown. |
Other listsPOLIS Staff and PhD Student Colloquium Cambridge Global Food Security Quastel Midgen LLP PresentationOther talks"Epigenetic studies in Alzheimer's disease" Stokes-Smoluchowski-Einstein-Langevin theory for active colloidal suspensions High-Dimensional Collocation for Lognormal Diffusion Problems Development of a Broadly-Neutralising Vaccine against Blood-Stage P. falciparum Malaria ‘Class-work’ in the elite institutions of higher education Single Cell Seminars (August) Cambridge-Lausanne Workshop 2018 - Day 1 Atiyah Floer conjecture The role of myosin VI in connexin 43 gap junction accretion Respiratory Problems |