PCToolkit: A Unified Plug-and-Play Prompt Compression Toolkit of Large Language Models

Abstract

Prompt engineering enables Large Language Models (LLMs) to perform a variety of tasks. However, lengthy prompts significantly increase computational complexity and economic costs. To address this issue, prompt compression reduces prompt length while maintaining LLM response quality. To support rapid implementation and standardization, we present the Prompt Compression Toolkit (PCToolkit), a unified plug-and-play framework for LLM prompt compression. PCToolkit integrates state-of-the-art compression algorithms, benchmark datasets, and evaluation metrics, enabling systematic performance analysis. Its modular architecture simplifies customization, offering portable interfaces for seamless incorporation of new datasets, metrics, and compression methods. Our code is available at https://github.com/3DAgentWorld/Toolkit-for-Prompt-Compression. Our demo is at https://huggingface.co/spaces/CjangCjengh/Prompt-Compression-Toolbox.

Cite

Text

Zhang et al. "PCToolkit: A Unified Plug-and-Play Prompt Compression Toolkit of Large Language Models." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/1277

Markdown

[Zhang et al. "PCToolkit: A Unified Plug-and-Play Prompt Compression Toolkit of Large Language Models." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/zhang2025ijcai-pctoolkit/) doi:10.24963/IJCAI.2025/1277

BibTeX

@inproceedings{zhang2025ijcai-pctoolkit,
  title     = {{PCToolkit: A Unified Plug-and-Play Prompt Compression Toolkit of Large Language Models}},
  author    = {Zhang, Zheng and Li, Jinyi and Lan, Yihuai and Wang, Xiang and Wang, Hao},
  booktitle = {International Joint Conference on Artificial Intelligence},
  year      = {2025},
  pages     = {11127-11131},
  doi       = {10.24963/IJCAI.2025/1277},
  url       = {https://mlanthology.org/ijcai/2025/zhang2025ijcai-pctoolkit/}
}