Search
Skip to Search Results- 25White, Martha (Computing Science)
- 21Bowling, Michael (Computing Science)
- 3Schuurmans, Dale (Computing Science)
- 3White, Adam (Computing Science)
- 1Bellemare, Marc (Google Brain)
- 1Farahmand, Amir-massoud (Computer Science, University of Toronto)
- 17Reinforcement Learning
- 10Machine Learning
- 7Artificial Intelligence
- 4Machine learning
- 4Reinforcement learning
- 3Exploration
-
Spring 2016
Game theoretic solution concepts, such as Nash equilibrium strategies that are optimal against worst case opponents, provide guidance in finding desirable autonomous agent behaviour. In particular, we wish to approximate solutions to complex, dynamic tasks, such as negotiation or bidding in...
-
Fall 2015
Extensive-form games are a powerful framework for modeling sequential multi-agent interactions. In extensive-form games with imperfect information, Nash equilibria are generally used as a solution concept, but computing a Nash equilibrium can be intractable in large games. Instead, a variety of...
-
Spring 2024
Optimistic value estimates provide one mechanism for directed exploration in reinforcement learning (RL). The agent acts greedily with respect to an estimate of the value plus what can be seen as a value bonus. The value bonus can be learned by estimating a value function on reward bonuses,...
-
Fall 2019
In this thesis, we investigate different vector step-size adaptation approaches for continual, online prediction problems. Vanilla stochastic gradient descent can be considerably improved by scaling the update with a vector of appropriately chosen step-sizes. Many methods, including AdaGrad,...
-
Fall 2023
Oftentimes, machine learning applications using neural networks involve solving discrete optimization problems, such as in pruning, parameter-isolation-based continual learning and training of binary networks. Still, these discrete problems are combinatorial in nature and are also not amenable to...