Is GPT-3 All You Need for Machine Learning for Chemistry?

Abstract

Pre-trained large language models (LLMs) are a powerful platform for building custom models for various applications. They have also found success in chemistry, but typically need to be pre-trained on large chemistry datasets such as reaction databases or protein sequences. In this work, we analyze whether one of the largest pre-trained LLMs, GPT-3, can be directly used for chemistry applications by fine-tuning on only a few data points from a chemistry dataset, i.e., without pre-training on a chemistry-specific dataset. We show that GPT-3 can achieve performance competing with baselines on three case studies (polymers, metal-organic frameworks, photoswitches) with representations as simple as the chemical name in both classification and regression settings. Moreover, we demonstrate that GPT-3 can also be fine-tuned for use in inverse design tasks, i.e., to generate a molecule that has properties as specified in a prompt.

Cite

Text

Jablonka et al. "Is GPT-3 All You Need for Machine Learning for Chemistry?." NeurIPS 2022 Workshops: AI4Mat, 2022.

Markdown

[Jablonka et al. "Is GPT-3 All You Need for Machine Learning for Chemistry?." NeurIPS 2022 Workshops: AI4Mat, 2022.](https://mlanthology.org/neuripsw/2022/jablonka2022neuripsw-gpt3/)

BibTeX

@inproceedings{jablonka2022neuripsw-gpt3,
  title     = {{Is GPT-3 All You Need for Machine Learning for Chemistry?}},
  author    = {Jablonka, Kevin Maik and Schwaller, Philippe and Smit, Berend},
  booktitle = {NeurIPS 2022 Workshops: AI4Mat},
  year      = {2022},
  url       = {https://mlanthology.org/neuripsw/2022/jablonka2022neuripsw-gpt3/}
}