Relational Embedding for Few-Shot Classification

Abstract

We propose to address the problem of few-shot classification by meta-learning "what to observe" and "where to attend" in a relational perspective. Our method leverages relational patterns within and between images via self-correlational representation (SCR) and cross-correlational attention (CCA). Within each image, the SCR module transforms a base feature map into a self-correlation tensor and learns to extract structural patterns from the tensor. Between the images, the CCA module computes cross-correlation between two image representations and learns to produce co-attention between them. Our Relational Embedding Network (RENet) combines the two relational modules to learn relational embedding in an end-to-end manner. In experimental evaluation, it achieves consistent improvements over state-of-the-art methods on four widely used few-shot classification benchmarks of miniImageNet, tieredImageNet, CUB-200-2011, and CIFAR-FS.

Cite

Text

Kang et al. "Relational Embedding for Few-Shot Classification." International Conference on Computer Vision, 2021. doi:10.1109/ICCV48922.2021.00870

Markdown

[Kang et al. "Relational Embedding for Few-Shot Classification." International Conference on Computer Vision, 2021.](https://mlanthology.org/iccv/2021/kang2021iccv-relational/) doi:10.1109/ICCV48922.2021.00870

BibTeX

@inproceedings{kang2021iccv-relational,
  title     = {{Relational Embedding for Few-Shot Classification}},
  author    = {Kang, Dahyun and Kwon, Heeseung and Min, Juhong and Cho, Minsu},
  booktitle = {International Conference on Computer Vision},
  year      = {2021},
  pages     = {8822-8833},
  doi       = {10.1109/ICCV48922.2021.00870},
  url       = {https://mlanthology.org/iccv/2021/kang2021iccv-relational/}
}