📍 Edmonton, Canada 🇨🇦
🔗 https://webdocs.cs.ualberta.ca/~machado/
🗓️ Joined November, 2024
― N.K. Jemisin, The Stone Sky
― N.K. Jemisin, The Stone Sky
(Originally published at TMLR)
- Deep RL track (Thu): AGaLiTe: Approximate Gated Linear Transformers for Online Reinforcement Learning by S. Pramanik
(Originally published at TMLR)
- Deep RL track (Thu): AGaLiTe: Approximate Gated Linear Transformers for Online Reinforcement Learning by S. Pramanik
(These are great papers!)
- Deep RL track (Thu): Deep Reinforcement Learning with Gradient Eligibility Traces by E. Elelimy
- Foundations track (Fri): An Analysis of Action-Value Temporal-Difference Methods That Learn State Values by B. Daley and P. Nagarajan
(These are great papers!)
- Deep RL track (Thu): Deep Reinforcement Learning with Gradient Eligibility Traces by E. Elelimy
- Foundations track (Fri): An Analysis of Action-Value Temporal-Difference Methods That Learn State Values by B. Daley and P. Nagarajan
Inductive Biases in RL
sites.google.com/view/ibrl-wo...
- A Study of Value-Aware Eigenoptions by H. Kotamreddy
Inductive Biases in RL
sites.google.com/view/ibrl-wo...
- A Study of Value-Aware Eigenoptions by H. Kotamreddy
RL Beyond Rewards
rlbrew2-workshop.github.io
- Tue 11:59 (spotlight talk): Towards An Option Basis To Optimize All Rewards by S. Chandrasekar
- The World Is Bigger: A Computationally-Embedded Perspective on the Big World Hypothesis by A. Lewandowsi
RL Beyond Rewards
rlbrew2-workshop.github.io
- Tue 11:59 (spotlight talk): Towards An Option Basis To Optimize All Rewards by S. Chandrasekar
- The World Is Bigger: A Computationally-Embedded Perspective on the Big World Hypothesis by A. Lewandowsi
Repo: github.com/machado-rese...
Website: agarcl.github.io
Preprint: arxiv.org/abs/2505.18347
Repo: github.com/machado-rese...
Website: agarcl.github.io
Preprint: arxiv.org/abs/2505.18347
It is perhaps no surprise that the classic algorithms we considered couldn't really make much progress in the full game.
It is perhaps no surprise that the classic algorithms we considered couldn't really make much progress in the full game.
arxiv.org/abs/2303.07507
arxiv.org/abs/2303.07507
Again, here's the preprint by Tse et al.: arxiv.org/abs/2505.16217
Again, here's the preprint by Tse et al.: arxiv.org/abs/2505.16217