Image-Embodied Knowledge Representation Learning

Abstract

Entity images could provide significant visual information for knowledge representation learning. Most conventional methods learn knowledge representations merely from structured triples, ignoring rich visual information extracted from entity images. In this paper, we propose a novel Image-embodied Knowledge Representation Learning model (IKRL), where knowledge representations are learned with both triple facts and images. More specifically, we first construct representations for all images of an entity with a neural image encoder. These image representations are then integrated into an aggregated image-based representation via an attention-based method. We evaluate our IKRL models on knowledge graph completion and triple classification. Experimental results demonstrate that our models outperform all baselines on both tasks, which indicates the significance of visual information for knowledge representations and the capability of our models in learning knowledge representations with images.

Cite

Text

Xie et al. "Image-Embodied Knowledge Representation Learning." International Joint Conference on Artificial Intelligence, 2017. doi:10.24963/IJCAI.2017/438

Markdown

[Xie et al. "Image-Embodied Knowledge Representation Learning." International Joint Conference on Artificial Intelligence, 2017.](https://mlanthology.org/ijcai/2017/xie2017ijcai-image/) doi:10.24963/IJCAI.2017/438

BibTeX

@inproceedings{xie2017ijcai-image,
  title     = {{Image-Embodied Knowledge Representation Learning}},
  author    = {Xie, Ruobing and Liu, Zhiyuan and Luan, Huanbo and Sun, Maosong},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2017},
  pages     = {3140-3146},
  doi       = {10.24963/IJCAI.2017/438},
  url       = {https://mlanthology.org/ijcai/2017/xie2017ijcai-image/}
}