This decommissioned ERA site remains active temporarily to support our final migration steps to https://ualberta.scholaris.ca, ERA's new home. All new collections and items, including Spring 2025 theses, are at that site. For assistance, please contact erahelp@ualberta.ca.
Search
Skip to Search Results- 101Reinforcement Learning
- 23Machine Learning
- 12Artificial Intelligence
- 6Transfer Learning
- 5Planning
- 5Representation Learning
- 91Graduate and Postdoctoral Studies (GPS), Faculty of
- 91Graduate and Postdoctoral Studies (GPS), Faculty of/Theses and Dissertations
- 5Computing Science, Department of
- 5Computing Science, Department of/Technical Reports (Computing Science)
- 3WISEST Summer Research Program
- 3WISEST Summer Research Program/WISEST Research Posters
-
Fall 2023
With the increasing complexity and capacity of modern Field-Programmable Gate Arrays (FPGAs), there is a growing demand for efficient FPGA computer-aided design (CAD) tools, particularly at the placement stage. While some previous works, such as RLPlace, have explored the efficacy of single-state...
-
Spring 2022
Policy gradient (PG) estimators are ineffective in dealing with softmax policies that are sub-optimally saturated, which refers to the situation when the policy concentrates its probability mass on sub-optimal actions. Sub-optimal policy saturation may arise from a bad policy initialization or a...
-
Automated Coordination of Distributed Energy Resources using Local Energy Markets and Reinforcement Learning
DownloadFall 2024
The conventional unidirectional model of the electricity grid operations is no longer sufficient. The continued proliferation of distributed energy resources and the resultant surge in net load variability at the grid edge necessitates deploying adequate demand response methods. This thesis...
-
Fall 2024
The sensitivity of reinforcement learning algorithm performance to hyperparameter choices poses a significant hurdle to the deployment of these algorithms in the real-world, where sampling can be limited by speed, safety, or other system constraints. To mitigate this, one approach is to learn a...
-
Fall 2023
In reinforcement learning (RL), agents learn to maximize a reward signal using nothing but observations from the environment as input to their decision making processes. Whether the agent is simple, consisting of only a policy that maps observations to actions, or complex, containing auxiliary...
-
Chasing Hallucinated Value: A Pitfall of Dyna Style Algorithms with Imperfect Environment Models
DownloadSpring 2020
In Dyna style algorithms, reinforcement learning (RL) agents use a model of the environment to generate simulated experience. By updating on this simulated experience, Dyna style algorithms allow agents to potentially learn control policies in fewer environment interactions than agents that use...
-
Fall 2023
Off-policy policy evaluation has been a critical and challenging problem in reinforcement learning, and Temporal-Difference (TD) learning is one of the most important approaches for addressing it. There has been significant interest in searching for off-policy TD algorithms which find the same...
-
Fall 2021
Learning auxiliary tasks, such as multiple predictions about the world, can provide many benets to reinforcement learning systems. A variety of off-policy learning algorithms have been developed to learn such predictions, but as yet there is little work on how to adapt the behavior to gather...
-
Fall 2023
Multilevel action selection is a reinforcement learning technique in which an action is broken into two parts, the type and the parameters. When using multilevel action selection in reinforcement learning, one must break the action space into multiple subsets. These subsets are typically disjoint...