Richtarik, Peter
161 publications
ICML
2025
ATA: Adaptive Task Allocation for Efficient Resource Management in Distributed Machine Learning
ECML-PKDD
2025
Collaborative Value Function Estimation Under Model Mismatch: A Federated Temporal Difference Analysis
TMLR
2025
Explicit Personalization and Local Training: Double Communication Acceleration in Federated Learning
TMLR
2025
GradSkip: Communication-Accelerated Local Gradient Methods with Better Computational Complexity
NeurIPS
2025
Local Curvature Descent: Squeezing More Curvature Out of Standard and Polyak Gradient Descent
ICLR
2025
Methods for Convex $(L_0,L_1)$-Smooth Optimization: Clipping, Acceleration, and Adaptivity
NeurIPS
2025
Second-Order Optimization Under Heavy-Tailed Noise: Hessian Clipping and Sample Complexity Limits
NeurIPS
2024
Byzantine Robustness and Partial Participation Can Be Achieved at Once: Just CLIP Gradient Differences
NeurIPSW
2024
Cohort Squeeze: Beyond a Single Communication Round per Cohort in Cross-Device Federated Learning
AISTATS
2024
Communication Compression for Byzantine Robust Learning: New Efficient Algorithms and Improved Rates
ICLR
2024
FedP3: Federated Personalized and Privacy-Friendly Network Pruning Under Model Heterogeneity
NeurIPSW
2024
LoCoDL: Communication-Efficient Distributed Learning with Local Training and Compression
NeurIPSW
2024
Local Curvature Descent: Squeezing More Curvature Out of Standard and Polyak Gradient Descent
NeurIPS
2024
MicroAdam: Accurate Adaptive Optimization with Low Space Overhead and Provable Convergence
NeurIPSW
2024
MindFlayer: Efficient Asynchronous Parallel SGD in the Presence of Heterogeneous and Random Worker Compute Times
NeurIPSW
2024
SPAM: Stochastic Proximal Point Method with Momentum Variance Reduction for Nonconvex Cross-Device Federated Learning
AISTATS
2024
Understanding Progressive Training Through the Framework of Randomized Coordinate Descent
NeurIPS
2023
2Direction: Theoretically Faster Distributed Training with Bidirectional Communication Compression
AISTATS
2023
Catalyst Acceleration of Error Compensated Methods Leads to Better Communication Complexity
NeurIPSW
2023
Det-CGD: Compressed Gradient Descent with Matrix Stepsizes for Non-Convex Optimization
NeurIPS
2023
Optimal Time Complexities of Parallel Stochastic Optimization Methods Under a Fixed Computation Model
NeurIPSW
2023
TAMUNA: Doubly Accelerated Federated Learning with Local Training, Compression, and Partial Participation
AISTATS
2022
Basis Matters: Better Communication-Efficient Second Order Methods for Federated Learning
AISTATS
2022
FLIX: A Simple and Communication-Efficient Alternative to Local Methods in Federated Learning
NeurIPS
2022
A Damped Newton Method Achieves Global $\mathcal O \left(\frac{1}{k^2}\right)$ and Local Quadratic Convergence Rate
NeurIPS
2022
Accelerated Primal-Dual Gradient Method for Smooth and Convex-Concave Saddle-Point Problems with Bilinear Coupling
NeurIPS
2022
BEER: Fast $O(1/T)$ Rate for Decentralized Nonconvex Optimization with Communication Compression
NeurIPS
2022
Communication Acceleration of Local Gradient Methods via an Accelerated Primal-Dual Algorithm with an Inexact Prox
NeurIPS
2022
Distributed Methods with Compressed Communication for Solving Variational Inequalities, with Theoretical Guarantees
NeurIPS
2022
Theoretically Better and Numerically Faster Distributed Optimization with Smoothness-Aware Quantization Techniques
AISTATS
2021
A Linearly Convergent Algorithm for Decentralized Optimization: Sending Less Bits for Free!
NeurIPS
2021
CANITA: Faster Rates for Distributed Convex Optimization with Communication Compression
NeurIPSW
2021
FedMix: A Simple and Communication-Efficient Alternative to Local Methods in Federated Learning
AISTATS
2020
A Unified Theory of SGD: Variance Reduction, Sampling, Quantization and Coordinate Descent
ALT
2020
Don’t Jump Through Hoops and Remove Those Loops: SVRG and Katyusha Are Better Without the Outer Loop