Single Layers of Attention Suffice to Predict Protein Contacts
Abstract
The established approach to unsupervised protein contact prediction estimates coevolving positions using undirected graphical models. This approach trains a Potts model on a Multiple Sequence Alignment. Increasingly large Transformers are being pretrained on unlabeled, unaligned protein sequence databases but have demonstrated mixed results for downstream tasks, including contact prediction. We argue that attention is a principled model of protein interactions, grounded in real properties of protein family data. We introduce an energy-based attention layer, factored attention, and show that it achieves comparable performance to Potts models while sharing parameters both within and across families. We contrast factored attention with the Transformer to indicate that the Transformer leverages hierarchical signal in protein family databases not captured by our single-layer models. This raises the exciting possibility for the development of powerful structured models of protein family databases.
Cite
Text
Bhattacharya et al. "Single Layers of Attention Suffice to Predict Protein Contacts." ICLR 2021 Workshops: EBM, 2021.Markdown
[Bhattacharya et al. "Single Layers of Attention Suffice to Predict Protein Contacts." ICLR 2021 Workshops: EBM, 2021.](https://mlanthology.org/iclrw/2021/bhattacharya2021iclrw-single/)BibTeX
@inproceedings{bhattacharya2021iclrw-single,
title = {{Single Layers of Attention Suffice to Predict Protein Contacts}},
author = {Bhattacharya, Nick and Thomas, Neil and Rao, Roshan and Dauparas, Justas and Koo, Peter K and Baker, David and Song, Yun S. and Ovchinnikov, Sergey},
booktitle = {ICLR 2021 Workshops: EBM},
year = {2021},
url = {https://mlanthology.org/iclrw/2021/bhattacharya2021iclrw-single/}
}