CAR-Transformer: Cross-Attention Reinforcement Transformer for Cross-Lingual Summarization

Abstract

Cross-Lingual Summarization (CLS) involves generating a summary for a given document in another language. Most of the existing approaches adopt multi-task training and knowledge distillation, which increases the training cost and improves the performance of CLS tasks intuitively but unexplainably. In this work, we propose Cross-Attention Reinforcement (CAR) module and incorporate the module into the transformer backbone to formulate the CAR-Transformer. The CAR module formulates a pseudo summarization policy parameterized by the cross-attention weights reinforced by the ground-truth monolingual summary without introducing extra model parameters. Our approach demonstrates more consistent improvement across CLS tasks compared to traditional multi-task training methods and outperforms the fine-tuned vanilla mBART by 3.67 and the best-performing multi-task training approach by 1.48 in ROUGE-L F1 score on the WikiLingua Korean-to-English CLS task.

Cite

Text

Cai and Yuan. "CAR-Transformer: Cross-Attention Reinforcement Transformer for Cross-Lingual Summarization." AAAI Conference on Artificial Intelligence, 2024. doi:10.1609/AAAI.V38I16.29724

Markdown

[Cai and Yuan. "CAR-Transformer: Cross-Attention Reinforcement Transformer for Cross-Lingual Summarization." AAAI Conference on Artificial Intelligence, 2024.](https://mlanthology.org/aaai/2024/cai2024aaai-car/) doi:10.1609/AAAI.V38I16.29724

BibTeX

@inproceedings{cai2024aaai-car,
  title     = {{CAR-Transformer: Cross-Attention Reinforcement Transformer for Cross-Lingual Summarization}},
  author    = {Cai, Yuang and Yuan, Yuyu},
  booktitle = {AAAI Conference on Artificial Intelligence},
  year      = {2024},
  pages     = {17718-17726},
  doi       = {10.1609/AAAI.V38I16.29724},
  url       = {https://mlanthology.org/aaai/2024/cai2024aaai-car/}
}