A Distributional Approach to Controlled Text Generation
Abstract
We propose a Distributional Approach for addressing Controlled Text Generation from pre-trained Language Models (LM). This approach permits to specify, in a single formal framework, both “pointwise’” and “distributional” constraints over the target LM — to our knowledge, the first model with such generality —while minimizing KL divergence from the initial LM distribution. The optimal target distribution is then uniquely determined as an explicit EBM (Energy-BasedModel) representation. From that optimal representation, we then train a target controlled Autoregressive LM through an adaptive distributional variant of PolicyGradient. We conduct a first set of experiments over pointwise constraints showing the advantages of our approach over a set of baselines, in terms of obtaining a controlled LM balancing constraint satisfaction with divergence from the pretrained LM. We then perform experiments over distributional constraints, a unique feature of our approach, demonstrating its potential as a remedy to the problem of Bias in Language Models. Through an ablation study, we show the effectiveness of our adaptive technique for obtaining faster convergence. Code available at https://github.com/naver/gdc
Cite
Text
Khalifa et al. "A Distributional Approach to Controlled Text Generation." International Conference on Learning Representations, 2021.Markdown
[Khalifa et al. "A Distributional Approach to Controlled Text Generation." International Conference on Learning Representations, 2021.](https://mlanthology.org/iclr/2021/khalifa2021iclr-distributional/)BibTeX
@inproceedings{khalifa2021iclr-distributional,
title = {{A Distributional Approach to Controlled Text Generation}},
author = {Khalifa, Muhammad and Elsahar, Hady and Dymetman, Marc},
booktitle = {International Conference on Learning Representations},
year = {2021},
url = {https://mlanthology.org/iclr/2021/khalifa2021iclr-distributional/}
}