PCToolkit: A Unified Plug-and-Play Prompt Compression Toolkit of Large Language Models
Abstract
Prompt engineering enables Large Language Models (LLMs) to perform a variety of tasks. However, lengthy prompts significantly increase computational complexity and economic costs. To address this issue, prompt compression reduces prompt length while maintaining LLM response quality. To support rapid implementation and standardization, we present the Prompt Compression Toolkit (PCToolkit), a unified plug-and-play framework for LLM prompt compression. PCToolkit integrates state-of-the-art compression algorithms, benchmark datasets, and evaluation metrics, enabling systematic performance analysis. Its modular architecture simplifies customization, offering portable interfaces for seamless incorporation of new datasets, metrics, and compression methods. Our code is available at https://github.com/3DAgentWorld/Toolkit-for-Prompt-Compression. Our demo is at https://huggingface.co/spaces/CjangCjengh/Prompt-Compression-Toolbox.
Cite
Text
Zhang et al. "PCToolkit: A Unified Plug-and-Play Prompt Compression Toolkit of Large Language Models." International Joint Conference on Artificial Intelligence, 2025. doi:10.24963/IJCAI.2025/1277Markdown
[Zhang et al. "PCToolkit: A Unified Plug-and-Play Prompt Compression Toolkit of Large Language Models." International Joint Conference on Artificial Intelligence, 2025.](https://mlanthology.org/ijcai/2025/zhang2025ijcai-pctoolkit/) doi:10.24963/IJCAI.2025/1277BibTeX
@inproceedings{zhang2025ijcai-pctoolkit,
title = {{PCToolkit: A Unified Plug-and-Play Prompt Compression Toolkit of Large Language Models}},
author = {Zhang, Zheng and Li, Jinyi and Lan, Yihuai and Wang, Xiang and Wang, Hao},
booktitle = {International Joint Conference on Artificial Intelligence},
year = {2025},
pages = {11127-11131},
doi = {10.24963/IJCAI.2025/1277},
url = {https://mlanthology.org/ijcai/2025/zhang2025ijcai-pctoolkit/}
}