Search
Skip to Search Results- 21Bowling, Michael (Computing Science)
- 14Schuurmans, Dale (Computing Science)
- 5Szepesvari, Csaba (Computing Science)
- 2Greiner, Russell (Computing Science)
- 1Bellemare, Marc (Google Brain)
- 1Bowling, Mike (Computing Science)
- 9Reinforcement Learning
- 6Artificial Intelligence
- 6Machine Learning
- 6Machine learning
- 3Game Theory
- 2Abstractions
-
Spring 2010
In this work, we present a unified, general approach to variance reduction in agent evaluation using machine learning to minimize variance. Evaluating an agent's performance in a stochastic setting is necessary for agent development, scientific evaluation, and competitions. Traditionally,...
-
Fall 2009
For zero-sum games, we have efficient solution techniques. Unfortunately, there are interesting games that are too large to solve. Here, a popular approach is to solve an abstract game that models the original game. We assume that more accurate the abstract games result in stronger strategies....
-
Spring 2016
Monte Carlo methods are a simple, effective, and widely deployed way of approximating integrals that prove too challenging for deterministic approaches. This thesis presents a number of contributions to the field of adaptive Monte Carlo methods. That is, approaches that automatically adjust the...
-
Spring 2015
Much of the focus on finding good representations in reinforcement learning has been on learning complex non-linear predictors of value. Methods like policy gradient, that do not learn a value function and instead directly represent policy, often need fewer parameters to learn good policies....
-
Spring 2024
In model-based reinforcement learning, an agent can improve its policy by planning: learning from experience generated by a model. Search control is the problem of determining which starting state should be used to generate this experience. Given a limited planning budget, an agent should be...
-
Advances in Probabilistic Generative Models: Normalizing Flows, Multi-View Learning, and Linear Dynamical Systems
DownloadFall 2020
This thesis considers some aspects of generative models including my contributions in deep probabilistic generative architectures and linear dynamical systems. First, some advances in deep probabilistic generative models are contributed. Flow-based generative modelling is an emerging and highly...
-
Spring 2023
Reinforcement learning (RL) defines a general computational problem where the learner must learn to make good decisions through interactive experience. To be effective in solving this problem, the learner must be able to explore the environment, make accurate predictions about the future, and...
-
Fall 2013
Due to its wide application in various fields, clustering, as a fundamental unsupervised learning problem, has been intensively investigated over the past few decades. Unfortunately, standard clustering formulations are known to be computationally intractable. Although many convex relaxations of...
-
Chasing Hallucinated Value: A Pitfall of Dyna Style Algorithms with Imperfect Environment Models
DownloadSpring 2020
In Dyna style algorithms, reinforcement learning (RL) agents use a model of the environment to generate simulated experience. By updating on this simulated experience, Dyna style algorithms allow agents to potentially learn control policies in fewer environment interactions than agents that use...
-
Spring 2017
Most machine learning problems can be posed as solving a mathematical program that describes the structure of the prediction problem, usually expressed in terms of carefully chosen losses and regularizers. However, many machine learning problems yield mathematical programs that are not convex in...