DiffER: Categorical Diffusion Models for Chemical Retrosynthesis

Abstract

Methods for automatic chemical retrosynthesis have found recent success through the application of models traditionally built for natural language processing, primarily through transformer neural networks. These models have demonstrated significant ability to translate between the SMILES encodings of chemical products and reactants, but are constrained as a result of their autoregressive nature. We propose DiffER, an alternative template-free method for retrosynthesis prediction in the form of categorical diffusion, which allows the entire output SMILES sequence to be predicted in unison. We construct an ensemble of diffusion models which achieves state of the art performance for top-1 accuracy and competitive performance for top-3 and top-5 accuracy. We prove that \ours is a strong baseline for a new class of template-free model and is capable of learning a variety of synthetic techniques used in laboratory settings.

Cite

Text

Current et al. "DiffER: Categorical Diffusion Models for Chemical Retrosynthesis." NeurIPS 2024 Workshops: AIDrugX, 2024.

Markdown

[Current et al. "DiffER: Categorical Diffusion Models for Chemical Retrosynthesis." NeurIPS 2024 Workshops: AIDrugX, 2024.](https://mlanthology.org/neuripsw/2024/current2024neuripsw-differ/)

BibTeX

@inproceedings{current2024neuripsw-differ,
  title     = {{DiffER: Categorical Diffusion Models for Chemical Retrosynthesis}},
  author    = {Current, Sean and Chen, Ziqi and Ning, Xia and Parthasarathy, Srinivasan},
  booktitle = {NeurIPS 2024 Workshops: AIDrugX},
  year      = {2024},
  url       = {https://mlanthology.org/neuripsw/2024/current2024neuripsw-differ/}
}