Levine, Sergey
460 publications
CoRL
2025
AutoEval: Autonomous Evaluation of Generalist Robot Manipulation Policies in the Real World
ICLRW
2025
AutoEval: Autonomous Evaluation of Generalist Robot Manipulation Policies in the Real World
NeurIPS
2025
Derivative-Free Guidance in Continuous and Discrete Diffusion Models with Soft Value-Based Decoding
ICML
2025
Hi Robot: Open-Ended Instruction Following with Hierarchical Vision-Language-Action Models
NeurIPS
2025
Knowledge Insulating Vision-Language-Action Models: Train Fast, Run Fast, Generalize Better
ICML
2025
Proposer-Agent-Evaluator (PAE): Autonomous Skill Discovery for Foundation Model Internet Agents
CoRL
2025
Reflective Planning: Vision-Language Models for Multi-Stage Long-Horizon Robotic Manipulation
NeurIPS
2025
Temporal Representation Alignment: Successor Features Enable Emergent Compositionality in Robot Instruction Following
NeurIPS
2024
Bridging Model-Based Optimization and Generative Modeling via Conservative Fine-Tuning of Diffusion Models
NeurIPSW
2024
Derivative-Free Guidance in Continuous and Discrete Diffusion Models with Soft Value-Based Decoding
NeurIPS
2024
Designing Cell-Type-Specific Promoter Sequences Using Conservative Model-Based Optimization
NeurIPS
2024
DigiRL: Training In-the-Wild Device-Control Agents with Autonomous Reinforcement Learning
ICMLW
2024
DigiRL: Training In-the-Wild Device-Control Agents with Autonomous Reinforcement Learning
ICMLW
2024
DigiRL: Training In-the-Wild Device-Control Agents with Autonomous Reinforcement Learning
ICMLW
2024
DigiRL: Training In-the-Wild Device-Control Agents with Autonomous Reinforcement Learning
NeurIPSW
2024
Fine-Tuning Discrete Diffusion Models via Reward Optimization with Applications to DNA and Protein Design
NeurIPS
2024
Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learning
NeurIPS
2024
Inference via Interpolation: Contrastive Representations Provably Enable Planning and Inference
CoRL
2024
Mobility VLA: Multimodal Instruction Navigation with Long-Context VLMs and Topological Graphs
CoRL
2024
Scaling Cross-Embodied Learning: One Policy for Manipulation, Navigation, Locomotion and Aviation
CoRL
2023
Goal Representations for Instruction Following: A Semi-Supervised Language Interface to Control
ICLRW
2023
Latent Conservative Objective Models for Offline Data-Driven Crystal Structure Prediction
NeurIPSW
2022
A Connection Between One-Step Regularization and Critic Regularization in Reinforcement Learning
NeurIPSW
2022
A Connection Between One-Step Regularization and Critic Regularization in Reinforcement Learning
CoRL
2022
Don’t Start from Scratch: Leveraging Prior Data to Automate Robotic Reinforcement Learning
ICMLW
2022
Effective Offline RL Needs Going Beyond Pessimism: Representations and Distributional Shift
NeurIPS
2022
First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual Information Maximization
CoRL
2022
Generalization with Lossy Affordances: Leveraging Broad Offline Data for Learning Visuomotor Tasks
NeurIPS
2022
Object Representations as Fixed Points: Training Iterative Refinement Algorithms with Implicit Differentiation
NeurIPSW
2022
Pre-Training for Robots: Leveraging Diverse Multitask Data via Offline Reinforcement Learning
NeurIPSW
2022
Pre-Training for Robots: Leveraging Diverse Multitask Data via Offline Reinforcement Learning
NeurIPSW
2022
Simplifying Model-Based RL: Learning Representations, Latent-Space Models, and Policies with One Objective
NeurIPS
2022
Unpacking Reward Shaping: Understanding the Benefits of Reward Engineering on Sample Complexity
CoRL
2021
Fully Autonomous Real-World Reinforcement Learning with Applications to Mobile Manipulation
ICMLW
2021
Intrinsic Control of Variational Beliefs in Dynamic Partially-Observed Visual Environments
NeurIPS
2021
Replacing Rewards with Examples: Example-Based Policy Search via Recursive Classification
ICML
2021
Variational Empowerment as Representation Learning for Goal-Conditioned Reinforcement Learning
NeurIPS
2021
Which Mutual-Information Representation Learning Objectives Are Sufficient for Control?
NeurIPS
2021
Why Generalization in RL Is Difficult: Epistemic POMDPs and Implicit Partial Observability
ICML
2020
Decentralized Reinforcement Learning: Global Decision-Making via Local Economic Transactions
NeurIPS
2020
Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model
ICLR
2019
From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following
NeurIPS
2019
MCP: Learning Composable Hierarchical Control with Multiplicative Compositional Policies
CoRL
2019
Relay Policy Learning: Solving Long-Horizon Tasks via Imitation and Reinforcement Learning
CoRL
2018
Composable Action-Conditioned Predictors: Flexible Off-Policy Learning for Robot Navigation
ICML
2018
Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor
ICML
2018
Universal Planning Networks: Learning Generalizable Representations for Visuomotor Control
NeurIPS
2018
Variational Inverse Control with Events: A General Framework for Data-Driven Reward Definition
ICML
2017
Combining Model-Based and Model-Free Updates for Trajectory-Centric Reinforcement Learning