Liu, Shiwei
57 publications
NeurIPS
2025
GPAS: Accelerating Convergence of LLM Pretraining via Gradient-Preserving Activation Scaling
NeurIPS
2024
AlphaPruning: Using Heavy-Tailed Self Regularization Theory for Improved Layer-Wise Pruning of Large Language Models
NeurIPS
2024
E2ENet: Dynamic Sparse Feature Fusion for Accurate and Efficient 3D Medical Image Segmentation
NeurIPS
2024
Found in the Middle: How Language Models Use Long Contexts Better via Plug-and-Play Positional Encoding
ICML
2024
Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
ICLRW
2024
Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
ICML
2023
Instant Soup: Cheap Pruning Ensembles in a Single Pass Can Draw Lottery Tickets from Large Models
AAAI
2023
Lottery Pools: Winning More by Interpolating Tickets Without Increasing Training or Inference Cost
NeurIPS
2023
Predicting Mutational Effects on Protein-Protein Binding via a Side-Chain Diffusion Probabilistic Model
NeurIPS
2023
The Emergence of Essential Sparsity in Large Pre-Trained Models: The Weights That Matter