Grammar Prompting for Domain-Specific Language Generation with Large Language Models
Abstract
Large language models (LLMs) can learn to perform a wide range of natural language tasks from just a handful of in-context examples. However, for generating strings from highly structured languages (e.g., semantic parsing to complex domain-specific languages), it is challenging for the LLM to generalize from just a few exemplars. We propose \emph{grammar prompting}, a simple approach to enable LLMs to use external knowledge and domain-specific constraints, expressed through a grammar in Backus--Naur Form (BNF), during in-context learning. Grammar prompting augments each demonstration example with a specialized grammar that is minimally sufficient for generating the particular output example, where the specialized grammar is a subset of the full DSL grammar. For inference, the LLM first predicts a BNF grammar given a test input, and then generates the output according to the rules of the grammar. Experiments demonstrate that grammar prompting can enable LLMs to perform competitively on a diverse set of DSL generation tasks, including semantic parsing (SMCalFlow, Overnight, GeoQuery), PDDL planning, and SMILES-based molecule generation.
Cite
Text
Wang et al. "Grammar Prompting for Domain-Specific Language Generation with Large Language Models." Neural Information Processing Systems, 2023.Markdown
[Wang et al. "Grammar Prompting for Domain-Specific Language Generation with Large Language Models." Neural Information Processing Systems, 2023.](https://mlanthology.org/neurips/2023/wang2023neurips-grammar/)BibTeX
@inproceedings{wang2023neurips-grammar,
title = {{Grammar Prompting for Domain-Specific Language Generation with Large Language Models}},
author = {Wang, Bailin and Wang, Zi and Wang, Xuezhi and Cao, Yuan and Saurous, Rif A. and Kim, Yoon},
booktitle = {Neural Information Processing Systems},
year = {2023},
url = {https://mlanthology.org/neurips/2023/wang2023neurips-grammar/}
}