Hong, Mingyi
90 publications
ICLR
2025
Joint Reward and Policy Learning with Demonstrations and Human Feedback Improves Alignment
ICLRW
2025
Reinforcement Learning in Inference Time: A Perspective from Successive Policy Iterations
ICML
2025
RoSTE: An Efficient Quantization-Aware Supervised Fine-Tuning Approach for Large Language Models
NeurIPS
2024
DOPPLER: Differentially Private Optimizers with Low-Pass Filter for Privacy Noise Reduction
NeurIPS
2024
Defensive Unlearning with Adversarial Training for Robust Concept Erasure in Diffusion Models
NeurIPSW
2024
DiSK: Differentially Private Optimizer with Simplified Kalman Filter for Noise Reduction
ICML
2024
EMC$^2$: Efficient MCMC Negative Sampling for Contrastive Learning with Global Convergence
NeurIPS
2024
Getting More Juice Out of the SFT Data: Reward Learning from Human Demonstration Improves SFT for LLM Alignment
TMLR
2024
Hybrid Federated Learning for Feature & Sample Heterogeneity: Algorithms and Implementation
AISTATS
2024
Krylov Cubic Regularized Newton: A Subspace Second-Order Method with Dimension-Free Convergence Rate
NeurIPSW
2024
Learning Reward and Policy Jointly from Demonstration and Preference Improves Alignment
NeurIPS
2024
RAW: A Robust and Agile Plug-and-Play Watermark Framework for AI-Generated Images with Provable Guarantees
NeurIPS
2024
SLTrain: A Sparse Plus Low Rank Approach for Parameter and Memory Efficient Pretraining
ICML
2023
FedAvg Converges to Zero Training Loss Linearly for Overparameterized Multi-Layer Neural Networks
NeurIPS
2023
Selectivity Drives Productivity: Efficient Dataset Pruning for Enhanced Transfer Learning
NeurIPS
2022
A Stochastic Linearized Augmented Lagrangian Method for Decentralized Bilevel Optimization
ICML
2022
A Stochastic Multi-Rate Control Framework for Modeling Distributed Optimization Algorithms
NeurIPSW
2022
Building Large Machine Learning Models from Small Distributed Models: A Layer Matching Approach
ICLR
2022
Decentralized Learning for Overparameterized Problems: A Multi-Agent Kernel Approximation Approach
NeurIPS
2022
Inducing Equilibria via Incentives: Simultaneous Design-and-Play Ensures Global Convergence
ICML
2022
Revisiting and Advancing Fast Adversarial Training Through the Lens of Bi-Level Optimization
ICML
2022
Understanding Clipping for Federated Learning: Convergence and Client-Level Differential Privacy
AISTATS
2021
Finding First-Order Nash Equilibria of Zero-Sum Games with the Regularized Nikaido-Isoda Function
NeurIPS
2020
Distributed Training with Heterogeneous Data: Bridging Median- and Mean-Based Algorithms
NeurIPS
2020
Finding Second-Order Stationary Points Efficiently in Smooth Nonconvex Linearly Constrained Optimization Problems
NeurIPS
2019
Provably Global Convergence of Actor-Critic: A Case for Linear Quadratic Regulator with Ergodic Cost
AISTATS
2016
An Improved Convergence Analysis of Cyclic Block Coordinate Descent-Type Methods for Strongly Convex Minimization
NeurIPS
2016
NESTT: A Nonconvex Primal-Dual Splitting Method for Distributed and Stochastic Optimization