AMBROSIA: A Benchmark for Parsing Ambiguous Questions into Database Queries

Abstract

Practical semantic parsers are expected to understand user utterances and map them to executable programs, even when these are ambiguous. We introduce a new benchmark, AMBROSIA, which we hope will inform and inspire the development of text-to-SQL parsers capable of recognizing and interpreting ambiguous requests. Our dataset contains questions showcasing three different types of ambiguity (scope ambiguity, attachment ambiguity, and vagueness), their interpretations, and corresponding SQL queries. In each case, the ambiguity persists even when the database context is provided. This is achieved through a novel approach that involves controlled generation of databases from scratch. We benchmark various LLMs on AMBROSIA, revealing that even the most advanced models struggle to identify and interpret ambiguity in questions.

Cite

Text

Saparina and Lapata. "AMBROSIA: A Benchmark for Parsing Ambiguous Questions into Database Queries." Neural Information Processing Systems, 2024. doi:10.52202/079017-2876

Markdown

[Saparina and Lapata. "AMBROSIA: A Benchmark for Parsing Ambiguous Questions into Database Queries." Neural Information Processing Systems, 2024.](https://mlanthology.org/neurips/2024/saparina2024neurips-ambrosia/) doi:10.52202/079017-2876

BibTeX

@inproceedings{saparina2024neurips-ambrosia,
  title     = {{AMBROSIA: A Benchmark for Parsing Ambiguous Questions into Database Queries}},
  author    = {Saparina, Irina and Lapata, Mirella},
  booktitle = {Neural Information Processing Systems},
  year      = {2024},
  doi       = {10.52202/079017-2876},
  url       = {https://mlanthology.org/neurips/2024/saparina2024neurips-ambrosia/}
}