A Pareto-Optimal Compositional Energy-Based Model for Sampling and Optimization of Protein Sequences

Abstract

Deep generative models have emerged as a popular machine learning-based approach for inverse design problems in the life sciences. However, these problems often require sampling new designs that satisfy multiple properties of interest in addition to learning the data distribution. This multi-objective optimization becomes more challenging when properties are independent or orthogonal to each other. In this work, we propose a Pareto-compositional energy-based model (pcEBM), a framework that uses multiple gradient descent for sampling new designs that adhere to various constraints in optimizing distinct properties. We demonstrate its ability to learn non-convex Pareto fronts and generate sequences that simultaneously satisfy multiple desired properties across a series of real-world antibody design tasks.

Cite

Text

Tagasovska et al. "A Pareto-Optimal Compositional Energy-Based Model for Sampling and Optimization of Protein Sequences." NeurIPS 2022 Workshops: AI4Science, 2022.

Markdown

[Tagasovska et al. "A Pareto-Optimal Compositional Energy-Based Model for Sampling and Optimization of Protein Sequences." NeurIPS 2022 Workshops: AI4Science, 2022.](https://mlanthology.org/neuripsw/2022/tagasovska2022neuripsw-paretooptimal/)

BibTeX

@inproceedings{tagasovska2022neuripsw-paretooptimal,
  title     = {{A Pareto-Optimal Compositional Energy-Based Model for Sampling and Optimization of Protein Sequences}},
  author    = {Tagasovska, Natasa and Frey, Nathan C. and Loukas, Andreas and Hotzel, Isidro and Lafrance-Vanasse, Julien and Kelly, Ryan Lewis and Wu, Yan and Rajpal, Arvind and Bonneau, Richard and Cho, Kyunghyun and Ra, Stephen and Gligorijevic, Vladimir},
  booktitle = {NeurIPS 2022 Workshops: AI4Science},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/tagasovska2022neuripsw-paretooptimal/}
}