Search
Skip to Search Results- 3Reinforcement Learning
- 1Approximate Value/Policy Iteration
- 1Confidence Sets
- 1Constrained Markov Decision Process
- 1Error Propagation
- 1Learning theory
-
Spring 2013
In a discrete-time online control problem, a learner makes an effort to control the state of an initially unknown environment so as to minimize the sum of the losses he suffers, where the losses are assumed to depend on the individual state-transitions. Various models of control problems have...
-
Fall 2023
Many real-world tasks in fields such as robotics and control can be formulated as constrained Markov decision processes (CMDPs). In CMDPs, the objective is usually to optimize the return while ensuring some constraints being satisfied at the same time. The primal-dual approach is a common...
-
Fall 2011
This thesis studies the reinforcement learning and planning problems that are modeled by a discounted Markov Decision Process (MDP) with a large state space and finite action space. We follow the value-based approach in which a function approximator is used to estimate the optimal value function....