Publications
Mastering Atari, Go, chess and shogi by planning with a learned model. J. Schrittwieser, I. Antonoglou, T. Hubert, K. Simonyan, L. Sifre, S. Schmitt, A. Guez, E. Lockhart, D. Hassabis, T. Graepel, T. Lillicrap, D. Silver. Nature, 2020.
Expected Eligibility Traces. H. V. Hasselt, S. Madjiheurem, M. Hessel, D. Silver, A. Barreto, D. Borsa. AAAI 2021.
The Value-Improvement Path Towards Better Representations for Reinforcement Learning. W. Dabney, A. Barreto, M. Rowland, R. Dadashi, J. Quan, M. G. Bellemare, D. Silver. AAAI 2021.
Discovering Reinforcement Learning Algorithms. J. Oh, M. Hessel, W. M. Czarnecki, Z. Xu, H. V. Hasselt, S. Singh, D. Silver. NeurIPS 2020.
Meta-Gradient Reinforcement Learning with an Objective Discovered Online. Z. Xu, H. V. Hasselt, M. Hessel, J. Oh, S. Singh, D. Silver. NeurIPS 2020.
Self-Tuning Deep Reinforcement Learning. T. Zahavy, Z. Xu, V. Veeriah, M. Hessel, J. Oh, H. van Hasselt, D. Silver, S. Singh. NeurIPS 2020.
Value-driven Hindsight Modelling. A. Guez, F. Viola, T. Weber, L. Buesing, S. Kapturowski, D. Precup, D. Silver, N. Heess. NeurIPS 2020.
The Value Equivalence Principle for Model-Based Reinforcement Learning. C. Grimm, A. Barreto, S. Singh, D. Silver. NeurIPS 2020.
The Option Keyboard Combining Skills in Reinforcement Learning. A. Barreto, D. Borsa, S. Hou, G. Comanici, E. Aygün, P. Hamel, D. Toyama, S. Mourad, D. Silver, D. Precup. NeurIPS 2019.