Structural Learning with Amortized Inference
Abstract
Training a structured prediction model involves performing several loss-augmented inference steps. Over the lifetime of the training, many of these inference problems, although different, share the same solution. We propose AI-DCD, an Amortized Inference framework for Dual Coordinate Descent method, an approximate learning algorithm, that accelerates the training process by exploiting this redundancy of solutions, without compromising the performance of the model. We show the efficacy of our method by training a structured SVM using dual coordinate descent for an entityrelation extraction task. Our method learns the same model as an exact training algorithm would, but call the inference engine only in 10% – 24% of the inference problems encountered during training. We observe similar gains on a multi-label classification task and with a Structured Perceptron model for the entity-relation task.
Cite
Text
Chang et al. "Structural Learning with Amortized Inference." AAAI Conference on Artificial Intelligence, 2015. doi:10.1609/AAAI.V29I1.9535Markdown
[Chang et al. "Structural Learning with Amortized Inference." AAAI Conference on Artificial Intelligence, 2015.](https://mlanthology.org/aaai/2015/chang2015aaai-structural/) doi:10.1609/AAAI.V29I1.9535BibTeX
@inproceedings{chang2015aaai-structural,
title = {{Structural Learning with Amortized Inference}},
author = {Chang, Kai-Wei and Upadhyay, Shyam and Kundu, Gourab and Roth, Dan},
booktitle = {AAAI Conference on Artificial Intelligence},
year = {2015},
pages = {2525-2531},
doi = {10.1609/AAAI.V29I1.9535},
url = {https://mlanthology.org/aaai/2015/chang2015aaai-structural/}
}