Analysis of the Attention in Tabular Language Models
Abstract
Recent transformer-based models for learning table representation have reported state-of-the-art results for different tasks such as table understanding, question answering and semantic parsing. The various proposed models use different architectures, specifically different attention mechanisms. In this paper, we analyze and compare the attention mechanisms used by two different tabular language models. By visualizing the attention maps of the models, we shed a light on the different patterns that the models exhibit. With our analysis on the aggregate attention over two tabular datasets, we provide insights which might help towards building more efficient models tailored for table representation learning.
Cite
Text
Koleva et al. "Analysis of the Attention in Tabular Language Models." NeurIPS 2022 Workshops: TRL, 2022.Markdown
[Koleva et al. "Analysis of the Attention in Tabular Language Models." NeurIPS 2022 Workshops: TRL, 2022.](https://mlanthology.org/neuripsw/2022/koleva2022neuripsw-analysis/)BibTeX
@inproceedings{koleva2022neuripsw-analysis,
title = {{Analysis of the Attention in Tabular Language Models}},
author = {Koleva, Aneta and Ringsquandl, Martin and Tresp, Volker},
booktitle = {NeurIPS 2022 Workshops: TRL},
year = {2022},
url = {https://mlanthology.org/neuripsw/2022/koleva2022neuripsw-analysis/}
}