Sabach, Shoham

17 publications

TMLR 2026 $\texttt{C2-DPO}$: Constrained Controlled Direct Preference Optimization Kavosh Asadi, Xingzi Xu, Julien Han, Ege Beyazit, Idan Pipano, Dominique Perrault-Joncas, Shoham Sabach, Mohammad Ghavamzadeh, Karim Bouyarmane
TMLR 2025 A Proximal Operator for Inducing 2:4-Sparsity Jonas M. Kübler, Yu-Xiang Wang, Shoham Sabach, Navid Ansari, Matthäus Kleindessner, Kailash Budhathoki, Volkan Cevher, George Karypis
ICML 2025 Comparing Few to Rank Many: Active Human Preference Learning Using Randomized Frank-Wolfe Method Kiran Koshy Thekumparampil, Gaurush Hiranandani, Kousha Kalantari, Shoham Sabach, Branislav Kveton
ICLRW 2025 Data-Efficient Supervised Fine-Tuning of Language Models Using Optimal Design Rohan Deb, Kiran Koshy Thekumparampil, Kousha Kalantari, Gaurush Hiranandani, Shoham Sabach, Branislav Kveton
ICML 2025 FisherSFT: Data-Efficient Supervised Fine-Tuning of Language Models Using Information Gain Rohan Deb, Kiran Koshy Thekumparampil, Kousha Kalantari, Gaurush Hiranandani, Shoham Sabach, Branislav Kveton
CPAL 2025 MoXCo: How I Learned to Stop Exploring and Love My Local Minima? Esha Singh, Shoham Sabach, Yu-Xiang Wang
ICML 2025 Proxsparse: Regularized Learning of Semi-Structured Sparsity Masks for Pretrained LLMs Hongyi Liu, Rajarshi Saha, Zhen Jia, Youngsuk Park, Jiaji Huang, Shoham Sabach, Yu-Xiang Wang, George Karypis
TMLR 2025 νSAM: Memory-Efficient Sharpness-Aware Minimization via Nuclear Norm Constraints Thomas Pethick, Parameswaran Raman, Lenon Minorics, Mingyi Hong, Shoham Sabach, Volkan Cevher
AISTATS 2024 Krylov Cubic Regularized Newton: A Subspace Second-Order Method with Dimension-Free Convergence Rate Ruichen Jiang, Parameswaran Raman, Shoham Sabach, Aryan Mokhtari, Mingyi Hong, Volkan Cevher
ICML 2024 Learning the Target Network in Function Space Kavosh Asadi, Yao Liu, Shoham Sabach, Ming Yin, Rasool Fakoor
ICML 2024 MADA: Meta-Adaptive Optimizers Through Hyper-Gradient Descent Kaan Ozkara, Can Karakus, Parameswaran Raman, Mingyi Hong, Shoham Sabach, Branislav Kveton, Volkan Cevher
ICLR 2024 TAIL: Task-Specific Adapters for Imitation Learning with Large Pretrained Models Zuxin Liu, Jesse Zhang, Kavosh Asadi, Yao Liu, Ding Zhao, Shoham Sabach, Rasool Fakoor
AISTATS 2023 Faster Projection-Free Augmented Lagrangian Methods via Weak Proximal Oracle Dan Garber, Tsur Livney, Shoham Sabach
NeurIPSW 2023 MoXCo:How I Learned to Stop Exploring and Love My Local Minima? Esha Singh, Shoham Sabach, Yu-Xiang Wang
NeurIPS 2023 Resetting the Optimizer in Deep RL: An Empirical Study Kavosh Asadi, Rasool Fakoor, Shoham Sabach
NeurIPSW 2023 TAIL: Task-Specific Adapters for Imitation Learning with Large Pretrained Models Zuxin Liu, Jesse Zhang, Kavosh Asadi, Yao Liu, Ding Zhao, Shoham Sabach, Rasool Fakoor
NeurIPS 2023 TD Convergence: An Optimization Perspective Kavosh Asadi, Shoham Sabach, Yao Liu, Omer Gottesman, Rasool Fakoor