This decommissioned ERA site remains active temporarily to support our final migration steps to https://ualberta.scholaris.ca, ERA's new home. All new collections and items, including Spring 2025 theses, are at that site. For assistance, please contact erahelp@ualberta.ca.
Search
Skip to Search Results- 2Approximate dynamic programming
- 2Function approximation
- 2Temporal difference learning
- 2Two-timescale stochastic approximation
- 1Active learning
- 1Actor-critic methods
-
Computationally effective optimization methods for complex process control and scheduling problems
DownloadFall 2011
Over the years, how to reduce the operational cost, raise the profit and enhance the operational safety attracts tremendous interests in the chemical and petroleum industry. Since the regulatory control strategy may not achieve such rigorous requirements, higher level process control activities,...
-
2009
Bhatnagar, Shalabh, Sutton, Richard, Ghavamzadeh, Mohammad, Lee, Mark
Technical report TR09-10. We present four new reinforcement learning algorithms based on actor-critic, function approximation, and natural gradient ideas, and we provide their convergence proofs. Actor-critic reinforcement learning methods are online approximations to policy iteration in which...