Evolving RL: Discovering New Activation Functions Using LLMs

Abstract

Deep Reinforcement Learning (DRL) has traditionally inherited activation functions from supervised learning, despite fundamental differences in learning dynamics and objectives. We present EvolveAct, a novel framework that leverages large language models and evolutionary search to automatically discover optimal activation functions for specific RL tasks. Our method combines genetic programming with code Large Language Models (LLMs) to explore a rich space of mathematical functions, optimizing for stability and performance in DRL training. Experimental results across multiple environments show that the discovered activation functions consistently outperform standard choices such as ReLU and TanH, improving final performance on the Minatar suite by 37.25% and 28.3% on the Brax suite on average. By jointly optimizing over multiple diverse environments, we discover activation functions that demonstrate strong generalization capabilities across different RL domains. This research provides a foundation for automating fundamental architectural choices in deep reinforcement learning systems.

Cite

Text

Nadimpalli et al. "Evolving RL: Discovering New Activation Functions Using LLMs." ICLR 2025 Workshops: AgenticAI, 2025.

Markdown

[Nadimpalli et al. "Evolving RL: Discovering New Activation Functions Using LLMs." ICLR 2025 Workshops: AgenticAI, 2025.](https://mlanthology.org/iclrw/2025/nadimpalli2025iclrw-evolving/)

BibTeX

@inproceedings{nadimpalli2025iclrw-evolving,
  title     = {{Evolving RL: Discovering New Activation Functions Using LLMs}},
  author    = {Nadimpalli, Kalyan Varma and Chirra, Shashank Reddy and Varakantham, Pradeep and Bauer, Stefan},
  booktitle = {ICLR 2025 Workshops: AgenticAI},
  year      = {2025},
  url       = {https://mlanthology.org/iclrw/2025/nadimpalli2025iclrw-evolving/}
}