Ofir Nachum
Ofir Nachum
Google Brain
Verified email at google.com
Title
Cited by
Cited by
Year
Data-Efficient Hierarchical Reinforcement Learning
O Nachum, S Gu, H Lee, S Levine
Advances in Neural Information Processing Systems, 2018
3872018
Learning to remember rare events
Ł Kaiser, O Nachum, A Roy, S Bengio
International Conference for Learning Representations, 2017
2872017
Bridging the gap between value and policy based reinforcement learning
O Nachum, M Norouzi, K Xu, D Schuurmans
arXiv preprint arXiv:1702.08892, 2017
2712017
Morphnet: Fast & simple resource-constrained structure learning of deep networks
A Gordon, E Eban, O Nachum, B Chen, H Wu, TJ Yang, E Choi
Proceedings of the IEEE conference on computer vision and pattern …, 2018
2412018
A Lyapunov-based Approach to Safe Reinforcement Learning
Y Chow, O Nachum, E Duenez-Guzman, M Ghavamzadeh
Advances in Neural Information Processing Systems, 2018
2282018
Behavior regularized offline reinforcement learning
Y Wu, G Tucker, O Nachum
arXiv preprint arXiv:1911.11361, 2019
1402019
D4rl: Datasets for deep data-driven reinforcement learning
J Fu, A Kumar, O Nachum, G Tucker, S Levine
arXiv preprint arXiv:2004.07219, 2020
1382020
Deep Reinforcement Learning for Vision-Based Robotic Grasping: A Simulated Comparative Evaluation of Off-Policy Methods
D Quillen, E Jang, O Nachum, C Finn, J Ibarz, S Levine
IEEE International Conference on Robotics and Automation, 2018
1372018
Dualdice: Behavior-agnostic estimation of discounted stationary distribution corrections
O Nachum, Y Chow, B Dai, L Li
arXiv preprint arXiv:1906.04733, 2019
1162019
Near-optimal representation learning for hierarchical reinforcement learning
O Nachum, S Gu, H Lee, S Levine
arXiv preprint arXiv:1810.01257, 2018
1042018
Deepmdp: Learning continuous latent space models for representation learning
C Gelada, S Kumar, J Buckman, O Nachum, MG Bellemare
International Conference on Machine Learning, 2170-2179, 2019
1012019
Identifying and correcting label bias in machine learning
H Jiang, O Nachum
International Conference on Artificial Intelligence and Statistics, 702-712, 2020
852020
Trust-pcl: An off-policy trust region method for continuous control
O Nachum, M Norouzi, K Xu, D Schuurmans
International Conference for Learning Representations, 2018
832018
Lyapunov-based safe policy optimization for continuous control
Y Chow, O Nachum, A Faust, E Duenez-Guzman, M Ghavamzadeh
arXiv preprint arXiv:1901.10031, 2019
762019
Algaedice: Policy gradient from arbitrary experience
O Nachum, B Dai, I Kostrikov, Y Chow, L Li, D Schuurmans
arXiv preprint arXiv:1912.02074, 2019
592019
RL Unplugged: Benchmarks for Offline Reinforcement Learning
Ç Gülçehre, Z Wang, A Novikov, T Le Paine, SG Colmenarejo, K Zolna, ...
46*2020
Imitation learning via off-policy distribution matching
I Kostrikov, O Nachum, J Tompson
arXiv preprint arXiv:1912.05032, 2019
402019
Why does hierarchy (sometimes) work so well in reinforcement learning?
O Nachum, H Tang, X Lu, S Gu, H Lee, S Levine
arXiv preprint arXiv:1909.10618, 2019
372019
Improving policy gradient by exploring under-appreciated rewards
O Nachum, M Norouzi, D Schuurmans
International Conference for Learning Representations, 2017
342017
Deployment-efficient reinforcement learning via model-based offline optimization
T Matsushima, H Furuta, Y Matsuo, O Nachum, S Gu
arXiv preprint arXiv:2006.03647, 2020
312020
The system can't perform the operation now. Try again later.
Articles 1–20