Patel, Kumar Kshitij

17 publications

NeurIPS 2025 Revisiting Consensus Error: A Fine-Grained Analysis of Local SGD Under Second-Order Data Heterogeneity Kumar Kshitij Patel, Ali Zindari, Sebastian U Stich, Lingxiao Wang
NeurIPSW 2024 Distributionally Robust Linear Regression with Block Lewis Weights Naren Sarayu Manoj, Kumar Kshitij Patel
ICLRW 2024 Efficient Private Federated Non-Convex Optimization with Shuffled Model Lingxiao Wang, Xingyu Zhou, Kumar Kshitij Patel, Lawrence Tang, Aadirupa Saha
ICMLW 2024 Grokking, Rank Minimization and Generalization in Deep Learning David Yunis, Kumar Kshitij Patel, Samuel Wheeler, Pedro Henrique Pamplona Savarese, Gal Vardi, Karen Livescu, Michael Maire, Matthew Walter
IJCAI 2024 Online Combinatorial Optimization with Group Fairness Constraints Negin Golrezaei, Rad Niazadeh, Kumar Kshitij Patel, Fransisca Susan
ICMLW 2024 Rank Minimization, Alignment and Weight Decay in Neural Networks David Yunis, Kumar Kshitij Patel, Samuel Wheeler, Pedro Henrique Pamplona Savarese, Gal Vardi, Karen Livescu, Michael Maire, Matthew Walter
COLT 2024 The Limits and Potentials of Local SGD for Distributed Heterogeneous Learning with Intermittent Communication Kumar Kshitij Patel, Margalit Glasgow, Ali Zindari, Lingxiao Wang, Sebastian U Stich, Ziheng Cheng, Nirmit Joshi, Nathan Srebro
ICML 2023 Federated Online and Bandit Convex Optimization Kumar Kshitij Patel, Lingxiao Wang, Aadirupa Saha, Nathan Srebro
ICMLW 2023 On the Still Unreasonable Effectiveness of Federated Averaging for Heterogeneous Distributed Learning Kumar Kshitij Patel, Margalit Glasgow, Lingxiao Wang, Nirmit Joshi, Nathan Srebro
NeurIPSW 2022 Distributed Online and Bandit Convex Optimization Kumar Kshitij Patel, Aadirupa Saha, Lingxiao Wang, Nathan Srebro
NeurIPSW 2022 On Convexity and Linear Mode Connectivity in Neural Networks David Yunis, Kumar Kshitij Patel, Pedro Henrique Pamplona Savarese, Gal Vardi, Jonathan Frankle, Matthew Walter, Karen Livescu, Michael Maire
NeurIPS 2022 Towards Optimal Communication Complexity in Distributed Non-Convex Optimization Kumar Kshitij Patel, Lingxiao Wang, Blake E Woodworth, Brian Bullins, Nati Srebro
ICLR 2020 Don't Use Large Mini-Batches, Use Local SGD Tao Lin, Sebastian U. Stich, Kumar Kshitij Patel, Martin Jaggi
ICML 2020 Is Local SGD Better than Minibatch SGD? Blake Woodworth, Kumar Kshitij Patel, Sebastian Stich, Zhen Dai, Brian Bullins, Brendan Mcmahan, Ohad Shamir, Nathan Srebro
NeurIPS 2020 Minibatch vs Local SGD for Heterogeneous Distributed Learning Blake E Woodworth, Kumar Kshitij Patel, Nati Srebro
NeurIPS 2019 Communication Trade-Offs for Local-SGD with Large Step Size Aymeric Dieuleveut, Kumar Kshitij Patel
MLJ 2019 Corruption-Tolerant Bandit Learning Sayash Kapoor, Kumar Kshitij Patel, Purushottam Kar