Folgen
Vivek Veeriah
Vivek Veeriah
Google DeepMind
Bestätigte E-Mail-Adresse bei google.com
Titel
Zitiert von
Zitiert von
Jahr
Differential recurrent neural networks for action recognition
V Veeriah, N Zhuang, GJ Qi
Proceedings of the IEEE international conference on computer vision, 4041-4049, 2015
6052015
Discovery of useful questions as auxiliary tasks
V Veeriah, M Hessel, Z Xu, J Rajendran, RL Lewis, J Oh, HP van Hasselt, ...
Advances in Neural Information Processing Systems 32, 2019
952019
A self-tuning actor-critic algorithm
T Zahavy, Z Xu, V Veeriah, M Hessel, J Oh, HP van Hasselt, D Silver, ...
Advances in neural information processing systems 33, 20913-20924, 2020
862020
Many-goals reinforcement learning
V Veeriah, J Oh, S Singh
arXiv preprint arXiv:1806.09605, 2018
572018
Discovery of options via meta-learned subgoals
V Veeriah, T Zahavy, M Hessel, Z Xu, J Oh, I Kemaev, HP van Hasselt, ...
Advances in Neural Information Processing Systems 34, 29861-29873, 2021
382021
Face valuing: Training user interfaces with facial expressions and reinforcement learning
V Veeriah, PM Pilarski, RS Sutton
arXiv preprint arXiv:1606.02807, 2016
292016
Robust hand gesture recognition algorithm for simple mouse control
V Veeriah, PL Swaminathan
International Journal of Computer and Communication Engineering 2 (2), 219, 2013
262013
Deep Learning Architecture with Dynamically Programmed Layers for Brain Connectome Prediction
V Veeriah J, R Durvasula, GJ Qi
ACM KDD 2015, 2015
212015
Reload: Reinforcement learning with optimistic ascent-descent for last-iterate convergence in constrained mdps
T Moskovitz, B O’Donoghue, V Veeriah, S Flennerhag, S Singh, T Zahavy
International Conference on Machine Learning, 25303-25336, 2023
172023
Tidbd: Adapting temporal-difference step-sizes through stochastic meta-descent
A Kearney, V Veeriah, JB Travnik, RS Sutton, PM Pilarski
arXiv preprint arXiv:1804.03334, 2018
172018
Diversifying ai: Towards creative chess with alphazero
T Zahavy, V Veeriah, S Hou, K Waugh, M Lai, E Leurent, N Tomasev, ...
arXiv preprint arXiv:2308.09175, 2023
152023
How Should an Agent Practice?
J Rajendran, R Lewis, V Veeriah, H Lee, S Singh
Proceedings of the AAAI Conference on Artificial Intelligence 34 (04), 5454-5461, 2020
112020
Learning feature relevance through step size adaptation in temporal-difference learning
A Kearney, V Veeriah, J Travnik, PM Pilarski, RS Sutton
arXiv preprint arXiv:1903.03252, 2019
112019
Forward actor-critic for nonlinear function approximation in reinforcement learning
V Veeriah, H van Seijen, RS Sutton
Proceedings of the 16th Conference on Autonomous Agents and MultiAgent …, 2017
112017
Crossprop: Learning representations by stochastic meta-gradient descent in neural networks
V Veeriah, S Zhang, RS Sutton
Machine Learning and Knowledge Discovery in Databases: European Conference …, 2017
92017
Learning state representations from random deep action-conditional predictions
Z Zheng, V Veeriah, R Vuorio, RL Lewis, S Singh
Advances in Neural Information Processing Systems 34, 23679-23691, 2021
62021
Grasp: Gradient-based affordance selection for planning
V Veeriah, Z Zheng, R Lewis, S Singh
arXiv preprint arXiv:2202.04772, 2022
42022
Learning options for action selection with meta-gradients in multi-task reinforcement learning
VVJ Veeraiah, TBZ Zahavy, M Hessel, Z Xu, J Oh, I Kemaev, ...
US Patent App. 17/918,365, 2023
12023
Discovery in Reinforcement Learning
V Veeriah
12022
Learning representations by stochastic meta-gradient descent in neural networks
V Veeriah, S Zhang, RS Sutton
arXiv preprint arXiv:1612.02879, 2016
12016
Das System kann den Vorgang jetzt nicht ausführen. Versuchen Sie es später erneut.
Artikel 1–20