Improving Precision in Language Models Learning from Invalid Samples
Abstract
Language Models are powerful generative tools capable of learning intricate patterns from vast amounts of unstructured data. Nevertheless, in domains that demand precision, such as science and engineering, the primary objective is to obtain an exact and accurate answer. Precision takes precedence in these contexts. In specialized tasks like chemical compound generation, the emphasis is on output accuracy rather than response diversity. Traditional self-refinement methods are ineffective for such domain-specific input/output pairs, unlike general language tasks. In this study, we introduce invalid2valid, a powerful and general post-processing mechanism that can significantly enhance precision in language models for input/output tasks spanning different domains and specialized applications.
Cite
Text
Larsen et al. "Improving Precision in Language Models Learning from Invalid Samples." NeurIPS 2023 Workshops: GenBio, 2023.Markdown
[Larsen et al. "Improving Precision in Language Models Learning from Invalid Samples." NeurIPS 2023 Workshops: GenBio, 2023.](https://mlanthology.org/neuripsw/2023/larsen2023neuripsw-improving/)BibTeX
@inproceedings{larsen2023neuripsw-improving,
title = {{Improving Precision in Language Models Learning from Invalid Samples}},
author = {Larsen, Niels and Giannone, Giorgio and Winther, Ole and Blin, Kai},
booktitle = {NeurIPS 2023 Workshops: GenBio},
year = {2023},
url = {https://mlanthology.org/neuripsw/2023/larsen2023neuripsw-improving/}
}