Distilling Structured Rationale from Large Language Models to Small Language Models for Abstractive Summarization

Abstract

Large Language Models (LLMs) have permeated various Natural Language Processing (NLP) tasks. For the summarization tasks, LLMs can generate well-structured rationales, which consist of Essential Aspects (EA), Associated Sentences (AS) and Triple Entity Relations (TER). These rationales guide smaller models (≤1B) to produce better summaries. However, their high deployment costs (≥70B), such as substantial storage space and high computing requirements, limit their utilization in resource-constrained environments. Furthermore, effectively distilling these structured rationales from LLMs into Small Language Models (SLMs) models remains a challenge. To address this, we propose the LLM-based Structured Rationale-guided Multi-view Weak-gated Fusion framework (LSR-MWF). The framework initially employs LLMs to dig structural rationales from a document, considering multiple viewpoints such as EA, AS, and TER. Then, it develop a multi-step summary generation evaluation strategy to select high-quality structured rationales. Subsequently, it aligns with these rationales using additional modules organized in a hierarchical structure. Finally, the framework integrates the features output by these modules with original abstractive model through a weak-gated mechanism. Experimental results on two publicly available CNN/DailyMail and XSum datasets show that our method improves the performance of the abstractive model, outperforming baselines by 11.2% and 5.8%, respectively. In addition, our method improves the interpretability of summary generation from the viewpoints of EA, AS and TER.

Cite

Text

Wang et al. "Distilling Structured Rationale from Large Language Models to Small Language Models for Abstractive Summarization." AAAI Conference on Artificial Intelligence, 2025. doi:10.1609/AAAI.V39I24.34727

Markdown

[Wang et al. "Distilling Structured Rationale from Large Language Models to Small Language Models for Abstractive Summarization." AAAI Conference on Artificial Intelligence, 2025.](https://mlanthology.org/aaai/2025/wang2025aaai-distilling/) doi:10.1609/AAAI.V39I24.34727

BibTeX

@inproceedings{wang2025aaai-distilling,
  title     = {{Distilling Structured Rationale from Large Language Models to Small Language Models for Abstractive Summarization}},
  author    = {Wang, Linyong and Wu, Lianwei and Song, Shaoqi and Wang, Yaxiong and Gao, Cuiyun and Wang, Kang},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {25389-25397},
  doi       = {10.1609/AAAI.V39I24.34727},
  url       = {https://mlanthology.org/aaai/2025/wang2025aaai-distilling/}
}