Maximum Entropy GFlowNets with Soft Q-Learning

Abstract

Generative Flow Networks (GFNs) have emerged as a powerful tool for sampling discrete objects from unnormalized distributions, offering a scalable alternative to Markov Chain Monte Carlo (MCMC) methods. While GFNs draw inspiration from maximum entropy reinforcement learning (RL), the connection between the two has largely been unclear and seemingly applicable only in specific cases. This paper addresses the connection by constructing an appropriate reward function, thereby establishing an exact relationship between GFNs and maximum entropy RL. This construction allows us to introduce maximum entropy GFNs, which achieve the maximum entropy attainable by GFNs without constraints on the state space, in contrast to GFNs with uniform backward policy.

Cite

Text

Mohammadpour et al. "Maximum Entropy GFlowNets with Soft Q-Learning." Artificial Intelligence and Statistics, 2024.

Markdown

[Mohammadpour et al. "Maximum Entropy GFlowNets with Soft Q-Learning." Artificial Intelligence and Statistics, 2024.](https://mlanthology.org/aistats/2024/mohammadpour2024aistats-maximum/)

BibTeX

@inproceedings{mohammadpour2024aistats-maximum,
  title     = {{Maximum Entropy GFlowNets with Soft Q-Learning}},
  author    = {Mohammadpour, Sobhan and Bengio, Emmanuel and Frejinger, Emma and Bacon, Pierre-Luc},
  booktitle = {Artificial Intelligence and Statistics},
  year      = {2024},
  pages     = {2593-2601},
  volume    = {238},
  url       = {https://mlanthology.org/aistats/2024/mohammadpour2024aistats-maximum/}
}