Hailes, Stephen

10 publications

ICLRW 2025 A Roadmap for Human-Agent Moral Alignment: Integrating Pre-Defined Intrinsic Rewards and Learned Reward Models Elizaveta Tennant, Stephen Hailes, Mirco Musolesi
ICLR 2025 Moral Alignment for LLM Agents Elizaveta Tennant, Stephen Hailes, Mirco Musolesi
ICLRW 2025 Moral Alignment for LLM Agents Elizaveta Tennant, Stephen Hailes, Mirco Musolesi
ICLRW 2025 Moral Intrinsic Rewards for Automated Alignment of LLM Agents Elizaveta Tennant, Stephen Hailes, Mirco Musolesi
AISTATS 2025 Partial Information Decomposition for Data Interpretability and Feature Selection Charles Westphal, Stephen Hailes, Mirco Musolesi
TMLR 2024 Graph Reinforcement Learning for Combinatorial Optimization: A Survey and Unifying Perspective Victor-Alexandru Darvariu, Stephen Hailes, Mirco Musolesi
IJCAI 2023 Modeling Moral Choices in Social Dilemmas with Multi-Agent Reinforcement Learning Elizaveta Tennant, Stephen Hailes, Mirco Musolesi
LoG 2022 Dynamic Network Reconfiguration for Entropy Maximization Using Deep Reinforcement Learning Christoffel Doorman, Victor-Alexandru Darvariu, Stephen Hailes, Mirco Musolesi
NeurIPS 2021 Solving Graph-Based Public Goods Games with Tree Search and Imitation Learning Victor-Alexandru Darvariu, Stephen Hailes, Mirco Musolesi
AAAI 2020 Partner Selection for the Emergence of Cooperation in Multi-Agent Systems Using Reinforcement Learning Nicolas Anastassacos, Stephen Hailes, Mirco Musolesi