Nachum, Ofir

61 publications

ICLR 2024 Multimodal Web Navigation with Instruction-Finetuned Foundation Models Hiroki Furuta, Kuang-Huei Lee, Ofir Nachum, Yutaka Matsuo, Aleksandra Faust, Shixiang Shane Gu, Izzeddin Gur
ICLR 2023 A Mixture-of-Expert Approach to RL-Based Dialogue Management Yinlam Chow, Azamat Tulepbergenov, Ofir Nachum, Dhawal Gupta, Moonkyung Ryu, Mohammad Ghavamzadeh, Craig Boutilier
CoRL 2023 Contrastive Value Learning: Implicit Models for Simple Offline RL Bogdan Mazoure, Benjamin Eysenbach, Ofir Nachum, Jonathan Tompson
ICLR 2023 Dichotomy of Control: Separating What You Can Control from What You Cannot Sherry Yang, Dale Schuurmans, Pieter Abbeel, Ofir Nachum
ICMLW 2023 In-Context Decision-Making from Supervised Pretraining Jonathan Lee, Annie Xie, Aldo Pacchiano, Yash Chandak, Chelsea Finn, Ofir Nachum, Emma Brunskill
ICLRW 2023 Instruction-Finetuned Foundation Models for Multimodal Web Navigation Hiroki Furuta, Ofir Nachum, Kuang-Huei Lee, Yutaka Matsuo, Shixiang Shane Gu, Izzeddin Gur
NeurIPS 2023 Inverse Dynamics Pretraining Learns Good Representations for Multitask Imitation David Brandfonbrener, Ofir Nachum, Joan Bruna
NeurIPS 2023 Learning Universal Policies via Text-Guided Video Generation Yilun Du, Sherry Yang, Bo Dai, Hanjun Dai, Ofir Nachum, Josh Tenenbaum, Dale Schuurmans, Pieter Abbeel
ICML 2023 Multi-Environment Pretraining Enables Transfer to Action Limited Datasets David Venuto, Sherry Yang, Pieter Abbeel, Doina Precup, Igor Mordatch, Ofir Nachum
ICLRW 2023 Multi-Environment Pretraining Enables Transfer to Action Limited Datasets David Venuto, Sherry Yang, Pieter Abbeel, Doina Precup, Igor Mordatch, Ofir Nachum
CoRL 2023 Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions Yevgen Chebotar, Quan Vuong, Karol Hausman, Fei Xia, Yao Lu, Alex Irpan, Aviral Kumar, Tianhe Yu, Alexander Herzog, Karl Pertsch, Keerthana Gopalakrishnan, Julian Ibarz, Ofir Nachum, Sumedh Anand Sontakke, Grecia Salazar, Huong T. Tran, Jodilyn Peralta, Clayton Tan, Deeksha Manjunath, Jaspiar Singh, Brianna Zitkovich, Tomas Jackson, Kanishka Rao, Chelsea Finn, Sergey Levine
NeurIPS 2023 Supervised Pretraining Can Learn In-Context Reinforcement Learning Jonathan Lee, Annie Xie, Aldo Pacchiano, Yash Chandak, Chelsea Finn, Ofir Nachum, Emma Brunskill
ICLRW 2023 Understanding HTML with Large Language Models Izzeddin Gur, Ofir Nachum, Yingjie Miao, Mustafa Safdari, Austin V Huang, Aakanksha Chowdhery, Sharan Narang, Noah Fiedel, Aleksandra Faust
AISTATS 2022 Offline Policy Selection Under Uncertainty Mengjiao Yang, Bo Dai, Ofir Nachum, George Tucker, Dale Schuurmans
NeurIPSW 2022 A Mixture-of-Expert Approach to RL-Based Dialogue Management Yinlam Chow, Azamat Tulepbergenov, Ofir Nachum, Dhawal Gupta, Moonkyung Ryu, Mohammad Ghavamzadeh, Craig Boutilier
NeurIPS 2022 Chain of Thought Imitation with Procedure Cloning Mengjiao Yang, Dale Schuurmans, Pieter Abbeel, Ofir Nachum
NeurIPSW 2022 Contrastive Value Learning: Implicit Models for Simple Offline RL Bogdan Mazoure, Benjamin Eysenbach, Ofir Nachum, Jonathan Tompson
NeurIPS 2022 Improving Zero-Shot Generalization in Offline Reinforcement Learning Using Generalized Similarity Functions Bogdan Mazoure, Ilya Kostrikov, Ofir Nachum, Jonathan J Tompson
ICML 2022 Model Selection in Batch Policy Optimization Jonathan Lee, George Tucker, Ofir Nachum, Bo Dai
NeurIPSW 2022 Multi-Environment Pretraining Enables Transfer to Action Limited Datasets David Venuto, Sherry Yang, Pieter Abbeel, Doina Precup, Igor Mordatch, Ofir Nachum
NeurIPS 2022 Multi-Game Decision Transformers Kuang-Huei Lee, Ofir Nachum, Mengjiao Yang, Lisa Lee, Daniel Freeman, Sergio Guadarrama, Ian Fischer, Winnie Xu, Eric Jang, Henryk Michalewski, Igor Mordatch
NeurIPS 2022 Oracle Inequalities for Model Selection in Offline Reinforcement Learning Jonathan N Lee, George Tucker, Ofir Nachum, Bo Dai, Emma Brunskill
ICLR 2022 Policy Gradients Incorporating the Future David Venuto, Elaine Lau, Doina Precup, Ofir Nachum
ICLR 2022 TRAIL: Near-Optimal Imitation Learning with Suboptimal Data Mengjiao Yang, Sergey Levine, Ofir Nachum
ICML 2022 Why Should I Trust You, Bellman? the Bellman Error Is a Poor Replacement for Value Error Scott Fujimoto, David Meger, Doina Precup, Ofir Nachum, Shixiang Shane Gu
NeurIPS 2022 Why so Pessimistic? Estimating Uncertainties for Offline RL Through Ensembles, and Why Their Independence Matters Kamyar Ghasemipour, Shixiang Gu, Ofir Nachum
ICLR 2021 Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization Michael R Zhang, Thomas Paine, Ofir Nachum, Cosmin Paduraru, George Tucker, Ziyu Wang, Mohammad Norouzi
ICLR 2021 Benchmarks for Deep Off-Policy Evaluation Justin Fu, Mohammad Norouzi, Ofir Nachum, George Tucker, Ziyu Wang, Alexander Novikov, Mengjiao Yang, Michael R Zhang, Yutian Chen, Aviral Kumar, Cosmin Paduraru, Sergey Levine, Thomas Paine
ICLR 2021 Deployment-Efficient Reinforcement Learning via Model-Based Offline Optimization Tatsuya Matsushima, Hiroki Furuta, Yutaka Matsuo, Ofir Nachum, Shixiang Gu
NeurIPS 2021 Near Optimal Policy Optimization via REPS Aldo Pacchiano, Jonathan N Lee, Peter L. Bartlett, Ofir Nachum
ICLR 2021 OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning Anurag Ajay, Aviral Kumar, Pulkit Agrawal, Sergey Levine, Ofir Nachum
NeurIPSW 2021 Offline Policy Selection Under Uncertainty Mengjiao Yang, Bo Dai, Ofir Nachum, George Tucker, Dale Schuurmans
ICML 2021 Offline Reinforcement Learning with Fisher Divergence Critic Regularization Ilya Kostrikov, Rob Fergus, Jonathan Tompson, Ofir Nachum
NeurIPSW 2021 Policy Gradients Incorporating the Future David Venuto, Elaine Lau, Doina Precup, Ofir Nachum
ICML 2021 Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning Hiroki Furuta, Tatsuya Matsushima, Tadashi Kozuno, Yutaka Matsuo, Sergey Levine, Ofir Nachum, Shixiang Shane Gu
NeurIPS 2021 Provable Representation Learning for Imitation with Contrastive Fourier Features Ofir Nachum, Mengjiao Yang
ICML 2021 Representation Matters: Offline Pretraining for Sequential Decision Making Mengjiao Yang, Ofir Nachum
ICMLW 2021 SparseDice: Imitation Learning for Temporally Sparse Data via Regularization Alberto Camacho, Izzeddin Gur, Marcin Lukasz Moczulski, Ofir Nachum, Aleksandra Faust
NeurIPSW 2021 Targeted Environment Design from Offline Data Izzeddin Gur, Ofir Nachum, Aleksandra Faust
IJCAI 2020 BRPO: Batch Residual Policy Optimization Sungryull Sohn, Yinlam Chow, Jayden Ooi, Ofir Nachum, Honglak Lee, Ed H. Chi, Craig Boutilier
NeurIPS 2020 CoinDICE: Off-Policy Confidence Interval Estimation Bo Dai, Ofir Nachum, Yinlam Chow, Lihong Li, Csaba Szepesvari, Dale Schuurmans
AISTATS 2020 Identifying and Correcting Label Bias in Machine Learning Heinrich Jiang, Ofir Nachum
ICLR 2020 Imitation Learning via Off-Policy Distribution Matching Ilya Kostrikov, Ofir Nachum, Jonathan Tompson
NeurIPS 2020 Off-Policy Evaluation via the Regularized Lagrangian Mengjiao Yang, Ofir Nachum, Bo Dai, Lihong Li, Dale Schuurmans
CoRL 2020 Safe Policy Learning for Continuous Control Yinlam Chow, Ofir Nachum, Aleksandra Faust, Edgar DueƱez-Guzman, Mohammad Ghavamzadeh
ICML 2019 DeepMDP: Learning Continuous Latent Space Models for Representation Learning Carles Gelada, Saurabh Kumar, Jacob Buckman, Ofir Nachum, Marc G. Bellemare
NeurIPS 2019 DualDICE: Behavior-Agnostic Estimation of Discounted Stationary Distribution Corrections Ofir Nachum, Yinlam Chow, Bo Dai, Lihong Li
ICMLW 2019 DualDICE: Efficient Estimation of Off-Policy Stationary Distribution Corrections Ofir Nachum, Yinlam Chow, Bo Dai, Lihong Li
ICMLW 2019 Lyapunov-Based Safe Policy Optimization for Continuous Control Yinlam Chow, Ofir Nachum, Aleksandra Faust, Edgar Duenez-Guzman, Mohammad Ghavamzadeh
CoRL 2019 Multi-Agent Manipulation via Locomotion Using Hierarchical Sim2Real Ofir Nachum, Michael Ahn, Hugo Ponte, Shixiang Gu, Vikash Kumar
ICLR 2019 Near-Optimal Representation Learning for Hierarchical Reinforcement Learning Ofir Nachum, Shixiang Gu, Honglak Lee, Sergey Levine
AISTATS 2019 Robustness Guarantees for Density Clustering Heinrich Jiang, Jennifer Jang, Ofir Nachum
ICLR 2019 The Laplacian in RL: Learning Representations with Efficient Approximations Yifan Wu, George Tucker, Ofir Nachum
NeurIPS 2018 A Lyapunov-Based Approach to Safe Reinforcement Learning Yinlam Chow, Ofir Nachum, Edgar Duenez-Guzman, Mohammad Ghavamzadeh
NeurIPS 2018 Data-Efficient Hierarchical Reinforcement Learning Ofir Nachum, Shixiang Gu, Honglak Lee, Sergey Levine
ICML 2018 Path Consistency Learning in Tsallis Entropy Regularized MDPs Yinlam Chow, Ofir Nachum, Mohammad Ghavamzadeh
ICML 2018 Smoothed Action Value Functions for Learning Gaussian Policies Ofir Nachum, Mohammad Norouzi, George Tucker, Dale Schuurmans
ICLR 2018 Trust-PCL: An Off-Policy Trust Region Method for Continuous Control Ofir Nachum, Mohammad Norouzi, Kelvin Xu, Dale Schuurmans
NeurIPS 2017 Bridging the Gap Between Value and Policy Based Reinforcement Learning Ofir Nachum, Mohammad Norouzi, Kelvin Xu, Dale Schuurmans
ICLR 2017 Improving Policy Gradient by Exploring Under-Appreciated Rewards Ofir Nachum, Mohammad Norouzi, Dale Schuurmans
ICLR 2017 Learning to Remember Rare Events Lukasz Kaiser, Ofir Nachum, Aurko Roy, Samy Bengio