Hao, Jianye
180 publications
NeurIPS
2025
COLA: Towards Efficient Multi-Objective Reinforcement Learning with Conflict Objective Regularization in Latent Space
NeurIPS
2025
CORE: Collaborative Optimization with Reinforcement Learning and Evolutionary Algorithm for Floorplanning
ICLR
2025
DistRL: An Asynchronous Distributed Reinforcement Learning Framework for On-Device Control Agent
NeurIPS
2025
Dynamic Configuration for Cutting Plane Separators via Reinforcement Learning on Incremental Graph
NeurIPS
2025
High-Performance Arithmetic Circuit Optimization via Differentiable Architecture Search
NeurIPS
2025
LogicTree: Improving Complex Reasoning of LLMs via Instantiated Multi-Step Synthetic Logical Data
NeurIPS
2025
OptiTree: Hierarchical Thoughts Generation with Tree Search for LLM Optimization Modeling
ICML
2025
STAR: Learning Diverse Robot Skill Abstractions Through Rotation-Augmented Vector Quantization
AAAI
2025
SWAMamba: A Sliding Window Attention Mamba Framework for Predicting Translation Elongation Rates
CVPR
2025
Spatial-Temporal Graph Diffusion Policy with Kinematic Modeling for Bimanual Robotic Manipulation
NeurIPS
2025
Succeed or Learn Slowly: Sample Efficient Off-Policy Reinforcement Learning for Mobile App Control
IJCAI
2025
The Graph's Apprentice: Teaching an LLM Low-Level Knowledge for Circuit Quality Estimation
NeurIPS
2025
Uncertainty-Quantified Rollout Policy Adaptation for Unlabelled Cross-Domain Video Temporal Grounding
ICML
2024
A Hierarchical Adaptive Multi-Task Reinforcement Learning Framework for Multiplier Circuit Design
NeurIPS
2024
CleanDiffuser: An Easy-to-Use Modularized Library for Diffusion Models in Decision Making
NeurIPSW
2024
DistRL: An Asynchronous Distributed Reinforcement Learning Framework for On-Device Control Agent
ICML
2024
EvoRainbow: Combining Improvements in Evolutionary Reinforcement Learning for Policy Search
NeurIPSW
2024
HuLE-Nav: Human-like Exploration for Zero-Shot Object Navigation via Vision-Language Models
ICML
2024
KISA: A Unified Keyframe Identifier and Skill Annotator for Long-Horizon Robotics Demonstrations
ICMLW
2024
Minimax Tree of Thoughts: Playing Two-Player Zero-Sum Sequential Games with Large Language Models
AAAI
2024
OVD-Explorer: Optimism Should Not Be the Sole Pursuit of Exploration in Noisy Environments
NeurIPS
2024
PERIA: Perceive, Reason, Imagine, Act via Holistic Language and Vision Planning for Manipulation
ICMLW
2024
SheetAgent: A Generalist Agent for Spreadsheet Reasoning and Manipulation via Large Language Models
NeurIPS
2024
The Ladder in Chaos: Improving Policy Learning by Harnessing the Parameter Evolving Path in a Low-Dimensional Space
NeurIPS
2024
Towards Next-Generation Logic Synthesis: A Scalable Neural Circuit Generation Framework
ICML
2024
Unlock the Cognitive Generalization of Deep Reinforcement Learning via Granular Ball Representation
ICMLW
2023
A Policy-Decoupled Method for High-Quality Data Augmentation in Offline Reinforcement Learning
ICMLW
2023
Boosting Off-Policy RL with Policy Representation and Policy-Extended Value Function Approximator
CVPR
2023
Co-Speech Gesture Synthesis by Reinforcement Learning with Contrastive Pre-Trained Rewards
ICLR
2023
EUCLID: Towards Efficient Unsupervised Reinforcement Learning with Multi-Choice Dynamics Model
AAAI
2023
Structure Aware Incremental Learning with Personalized Imitation Weights for Recommender Systems
ECML-PKDD
2022
Branch Ranking for Efficient Mixed-Integer Programming via Offline Ranking-Based Policy Learning
NeurIPS
2022
DOMINO: Decomposed Mutual Information Optimization for Generalized Context in Meta-Reinforcement Learning
NeurIPSW
2022
EUCLID: Towards Efficient Unsupervised Reinforcement Learning with Multi-Choice Dynamics Model
ICLR
2022
HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation
ICML
2022
PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration
NeurIPS
2022
Plan to Predict: Learning an Uncertainty-Foreseeing Model for Model-Based Reinforcement Learning
NeurIPSW
2022
Towards a Unified Policy Abstraction Theory and Representation Learning Approach in Markov Decision Processes
NeurIPS
2022
Transformer-Based Working Memory for Multiagent Reinforcement Learning with Action Parsing
NeurIPS
2021
Flattening Sharpness for Dynamic Gradient Projection Memory Benefits Continual Learning
NeurIPSW
2021
HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation
NeurIPSW
2021
OVD-Explorer: A General Information-Theoretic Exploration Approach for Reinforcement Learning
AAAI
2021
Towards Effective Context for Meta-Reinforcement Learning: An Approach Based on Contrastive Learning
IJCAI
2020
Generating Behavior-Diverse Game AIs with Evolutionary Multi-Objective Deep Reinforcement Learning