Latte: Transfering LLMs' Latent-Level Knowledge for Few-Shot Tabular Learning

Abstract

Few-shot tabular learning, in which machine learning models are trained with a limited amount of labeled data, provides a cost-effective approach to addressing real-world challenges. The advent of Large Language Models (LLMs) has sparked interest in leveraging their pre-trained knowledge for few-shot tabular learning. Despite promising results, existing approaches either rely on test-time knowledge extraction, which introduces undesirable latency, or text-level knowledge, which leads to unreliable feature engineering. To overcome these limitations, we propose Latte, a training-time knowledge extraction framework that transfers the latent prior knowledge within LLMs to optimize a more generalized downstream model. Latte enables general knowledge-guided downstream tabular learning, facilitating the weighted fusion of information across different feature values while reducing the risk of overfitting to limited labeled data. Furthermore, Latte is compatible with existing unsupervised pre-training paradigms and effectively utilizes available unlabeled samples to overcome the performance limitations imposed by an extremely small labeled dataset. Extensive experiments on various few-shot tabular learning benchmarks demonstrate the superior performance of Latte, establishing it as a state-of-the-art approach in this domain. Our code is available at https://github.com/ruxueshi/Latte.git.

Cite

Text

Shi et al. "Latte: Transfering LLMs' Latent-Level Knowledge for Few-Shot Tabular Learning." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/687

Markdown

[Shi et al. "Latte: Transfering LLMs' Latent-Level Knowledge for Few-Shot Tabular Learning." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/shi2025ijcai-latte/) doi:10.24963/IJCAI.2025/687

BibTeX

@inproceedings{shi2025ijcai-latte,
  title     = {{Latte: Transfering LLMs' Latent-Level Knowledge for Few-Shot Tabular Learning}},
  author    = {Shi, Ruxue and Gu, Hengrui and Ye, Hangting and Dai, Yiwei and Shen, Xu and Wang, Xin},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {6173-6181},
  doi       = {10.24963/IJCAI.2025/687},
  url       = {https://mlanthology.org/ijcai/2025/shi2025ijcai-latte/}
}