1 code implementation • 23 Jun 2023 • Shibhansh Dohare, J. Fernando Hernandez-Garcia, Parash Rahman, A. Rupam Mahmood, Richard S. Sutton
If deep-learning systems are applied in a continual learning setting, then it is well known that they may fail to remember earlier examples.
no code implementations • 9 Dec 2019 • J. Fernando Hernandez-Garcia, Richard S. Sutton
Sparse representations have been shown to be useful in deep reinforcement learning for mitigating catastrophic interference and improving the performance of agents in terms of cumulative reward.
1 code implementation • 22 Jan 2019 • J. Fernando Hernandez-Garcia, Richard S. Sutton
Our results show that (1) using off-policy correction can have an adverse effect on the performance of Sarsa and $Q(\sigma)$; (2) increasing the backup length $n$ consistently improved performance across all the different algorithms; and (3) the performance of Sarsa and $Q$-learning was more robust to the effect of the target network update frequency than the performance of Tree Backup, $Q(\sigma)$, and Retrace in this particular task.
no code implementations • 3 Mar 2017 • Kristopher De Asis, J. Fernando Hernandez-Garcia, G. Zacharias Holland, Richard S. Sutton
These methods are often studied in the one-step case, but they can be extended across multiple time steps to achieve better performance.