Binding Oracle: Fine-Tuning from Stability to Binding Free Energy
Abstract
The ability to predict changes in binding free energy (▵▵$G_{bind}$) for mutations at protein-protein interfaces (PPIs) is critical for the understanding genetic diseases and engineering novel protein-based therapeutics. Here, we present Binding Oracle: a structure-based graph transformer for predicting ▵▵$G_{bind}$ at PPIs. Binding Oracle fine-tunes Stability Oracle with Selective LoRA: a technique that synergizes layer selection via gradient norms with LoRA. Selective LoRA enables the identification and fine-tuning of the layers most critical for the downstream task, thus, regularizing against overfitting. Additionally, we present new training-test splits of mutational data from the SKEMPI2.0, Ab-Bind, and NABE databases that use a strict 30\% sequence similarity threshold to avoid data leakage during model evaluation. Binding Oracle, when trained with the Thermodynamic Permutations data augmentation technique , achieves SOTA on S487 without using any evolutionary auxiliary features. Our results empirically demonstrate how sparse fine-tuning techniques, such as Selective LoRA, can enable rapid domain adaptation in protein machine learning frameworks.
Cite
Text
Gong et al. "Binding Oracle: Fine-Tuning from Stability to Binding Free Energy." NeurIPS 2023 Workshops: GenBio, 2023.Markdown
[Gong et al. "Binding Oracle: Fine-Tuning from Stability to Binding Free Energy." NeurIPS 2023 Workshops: GenBio, 2023.](https://mlanthology.org/neuripsw/2023/gong2023neuripsw-binding/)BibTeX
@inproceedings{gong2023neuripsw-binding,
title = {{Binding Oracle: Fine-Tuning from Stability to Binding Free Energy}},
author = {Gong, Chengyue and Klivans, Adam and Wells, Jordan and Loy, James and Liu, Qiang and Dimakis, Alex and Diaz, Daniel},
booktitle = {NeurIPS 2023 Workshops: GenBio},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/gong2023neuripsw-binding/}
}