Liu, Xiang
29 publications
ICML
2025
Can Compressed LLMs Truly Act? an Empirical Evaluation of Agentic Capabilities in LLM Compression
NeurIPS
2025
ChunkKV: Semantic-Preserving KV Cache Compression for Efficient Long-Context LLM Inference
NeurIPS
2024
LISA: Layerwise Importance Sampling for Memory-Efficient Large Language Model Fine-Tuning