Search
Skip to Search Results- 22White, Martha (Computing Science)
- 20Bowling, Michael (Computing Science)
- 3Schuurmans, Dale (Computing Science)
- 3White, Adam (Computing Science)
- 1Bellemare, Marc (Google Brain)
- 1Farahmand, Amir-massoud (Computer Science, University of Toronto)
- 14Reinforcement Learning
- 8Machine Learning
- 7Artificial Intelligence
- 4Machine learning
- 3Game Theory
- 3Neural Networks
-
Spring 2022
A key problem in the theory of meta-learning is to understand how the task distributions influence transfer risk, the expected error of a meta-learner on a new task drawn from the unknown task distribution. In this work, focusing on fixed design linear regression with Gaussian noise and a...
-
Spring 2010
In this work, we present a unified, general approach to variance reduction in agent evaluation using machine learning to minimize variance. Evaluating an agent's performance in a stochastic setting is necessary for agent development, scientific evaluation, and competitions. Traditionally,...
-
Fall 2009
For zero-sum games, we have efficient solution techniques. Unfortunately, there are interesting games that are too large to solve. Here, a popular approach is to solve an abstract game that models the original game. We assume that more accurate the abstract games result in stronger strategies....
-
Spring 2016
Monte Carlo methods are a simple, effective, and widely deployed way of approximating integrals that prove too challenging for deterministic approaches. This thesis presents a number of contributions to the field of adaptive Monte Carlo methods. That is, approaches that automatically adjust the...
-
Spring 2015
Much of the focus on finding good representations in reinforcement learning has been on learning complex non-linear predictors of value. Methods like policy gradient, that do not learn a value function and instead directly represent policy, often need fewer parameters to learn good policies....
-
Spring 2020
Reinforcement Learning is a formalism for learning by trial and error. Unfortunately, trial and error can take a long time to find a solution if the agent does not efficiently explore the behaviours available to it. Moreover, how an agent ought to explore depends on the task that the agent is...
-
Spring 2022
Policy gradient (PG) estimators are ineffective in dealing with softmax policies that are sub-optimally saturated, which refers to the situation when the policy concentrates its probability mass on sub-optimal actions. Sub-optimal policy saturation may arise from a bad policy initialization or a...
-
Chasing Hallucinated Value: A Pitfall of Dyna Style Algorithms with Imperfect Environment Models
DownloadSpring 2020
In Dyna style algorithms, reinforcement learning (RL) agents use a model of the environment to generate simulated experience. By updating on this simulated experience, Dyna style algorithms allow agents to potentially learn control policies in fewer environment interactions than agents that use...
-
Fall 2023
Multilevel action selection is a reinforcement learning technique in which an action is broken into two parts, the type and the parameters. When using multilevel action selection in reinforcement learning, one must break the action space into multiple subsets. These subsets are typically disjoint...
-
Fall 2023
The problem of missing data is omnipresent in a wide range of real-world datasets. When learning and predicting on this data with neural networks, the typical strategy is to fill-in or complete these missing values in the dataset, called impute-then-regress. Much less common is to attempt to...