Skip to content
View dyth's full-sized avatar

Block or report dyth

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Please don't include any personal information such as legal names or email addresses. Maximum 100 characters, markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
dyth/README.md

David Yu-Tung Hui, 許宇同

I am interested in Deep Reinforcement Learning and its application in continuous-control tasks. My research focused on improving the optimization stability of off-policy gradient-based $Q$-learning algorithms over a range of tasks and hyperparameters.

I've written two works along this research direction:

  1. Stabilizing Q-Learning for Continuous Control
    David Yu-Tung Hui
    MSc Thesis, University of Montreal, 2022
    There are two topics in this thesis. First, I investigated the duality between maximizing entropy and maximizing likelihood in RL. Then, I showed that LayerNorm reduced divergence in $Q$-learning for high-dimensional continuous control.
    [.pdf] [Errata]

  2. Double Gumbel Q-Learning
    David Yu-Tung Hui, Aaron Courville, Pierre-Luc Bacon
    Spotlight at NeurIPS 2023
    We showed that using function approximation in $Q$-learning introduced two heteroscedastic Gumbel noise sources. Modeling these noise sources improved by up to $2\times$ the aggregate performance of a $Q$-learning algorithm at 1M timesteps over 33 continuous control tasks.
    [.pdf] [Reviews] [Poster (.png)] [5-min talk] [1-hour seminar] [Code (GitHub)] [Errata]

See my Google Scholar for more of my research.

Pinned Loading

  1. doublegum doublegum Public

    NeurIPS 2023 Spotlight

    Python 10 3