About Me
I am an assistant professor at Department of Computer Science, George Mason University. Before that I was a postdoc at Rafik B. Hariri Institute at Boston University, hosted by Francesco Orabona. I received my Ph.D. at Department of Computer Science, The University of Iowa in August 2020, under the advise of Tianbao Yang. Before that I studied at Institute of Natural Sciences and School of Mathematical Sciences at Shanghai Jiao Tong University. I have also spent time working at industrial research labs, such as IBM research AI and Alibaba DAMO Academy. Here is my Google Scholar Citations.
I am looking for selfmotivated PhD students (fullyfunded) with strong mathematical ablities and (or) programming skills to solve challenging machine learning problems elegantly with mathematical analysis and empirical studies. The main technical skills we need include mathematical optimization, statistical learning theory, algorithms, and deep learning. If you are interested, please drop me an email with your CV and transcript, and apply our PhD program here. Undergrad and graduate student visitors are also welcome. This link provides an overview of our fastgrowing GMU CS department.
I am looking for selfmotivated PhD students (fullyfunded) with strong mathematical ablities and (or) programming skills to solve challenging machine learning problems elegantly with mathematical analysis and empirical studies. The main technical skills we need include mathematical optimization, statistical learning theory, algorithms, and deep learning. If you are interested, please drop me an email with your CV and transcript, and apply our PhD program here. Undergrad and graduate student visitors are also welcome. This link provides an overview of our fastgrowing GMU CS department.
Research
My research interests are machine learning, optimization, learning theory, and deep learning. My goal is to design provably efficient algorithms for machine learning problems with strong empirical performance. In particular, I work on
Mathematical Optimization for Machine Learning: I focus on designing provably efficient optimization algorithms for machine (deep) learning problems, such as training language models, optimizing complex metrics (e.g., AUC maximization, Fmeasure optimization), hierarchical optimization (e.g., Generative Adversarial Nets), etc.
Statistical Learning Theory: I am interested in sample complexity and computational complexity for modern machine learning problems.
Largescale Distributed Learning: I design efficient scalable learning algorithms for distributed intelligence under various constraints (e.g., communication, privacy, etc.)
Machine Learning Applications: continual learning, model compression/quantization, autoML.
News
 (Apr 2023) I am happy to give an invited talk at Thomas Jefferson High School of Science and Technology about "Federated Learning: Algorithm Design and Applications". Here is the slides.
 (Mar 2023) I will be serving as an Area Chair for NeurIPS 2023.
 (Mar 2023) Glad to give an invited talk at IBM Almaden Research Center about new federated and adaptive optimization algorithms for deep learning with unbounded smooth landscape.
 (Feb 2023) Glad to give an invited talk at Google about new federated and adaptive optimization algorithms for deep learning with unbounded smooth landscape.
 (Jan 2023) One paper was accepted by ICLR 2023. Congratulations to my students!
 (Nov 2022) Two NeurIPS 2022 papers CommunicationEfficient Distributed Gradient Clipping and Bilevel Optimization were selected as spotlight presentations (5% acceptance rate).
 (Sep 2022) Three papers were accepted by NeurIPS 2022. Congratulations to my students and coauthors.
 (May 2022) One paper was accepted by ICML 2022. Congratulations to my students and coauthors.
 (Feb 2022) We will give a tutorial on CVPR 2022 about "Deep AUC Maximization". Here is the website.
 (Feb 2022) Two papers were accepted by ALT 2022.
 (Sep 2021) One paper was accepted by NeurIPS 2021.
 (Aug 2021) Joined in CS @ GMU as an assistant professor.
 (June 2021) The paper about nonconvexnonconcave minmax optimization has been accepted by JMLR.
 (Sep 2020) Two papers were accepted by NeurIPS 2020, with one Spotlight presentation (3% acceptance rate).
 (May 2020) One paper was accepted by ICML 2020.
 (Dec 2019) Two papers were accepted by ICLR 2020.
 (Sep 2018) Three papers were accepted by NeurIPS 2018.
 (May 2018) One paper was accepted by ICML 2018.
Recent Publications [Full List]

#: student author, *: equal contribution (alphabetical order)
 (New! )
EPISODE: Episodic Gradient Clipping with Periodic Resampled Corrections for Federated Learning with Heterogeneous Data
Michael Crawshaw^{#}, Yajie Bao^{#}, Mingrui Liu.
To appear in 11th International Conference on Learning Representations, 2023. (ICLR 2023)

A CommunicationEfficient Distributed Gradient Clipping Algorithm for Training Deep Neural Networks
Mingrui Liu, Zhenxun Zhuang, Yunwen Lei, Chunyang Liao.
In Advances in Neural Information Processing Systems 36, 2022. (NeurIPS 2022) (Spotlight, 5% acceptance rate)

Robustness to Unbounded Smoothness of Generalized SignSGD
Michael Crawshaw^{#*}, Mingrui Liu^{*}, Francesco Orabona^{*}, Wei Zhang^{*}, Zhenxun Zhuang^{*}.
In Advances in Neural Information Processing Systems 36, 2022. (NeurIPS 2022)

Will Bilevel Optimizers Benefit from Loops
Kaiyi Ji, Mingrui Liu, Yingbin Liang, Lei Ying.
In Advances in Neural Information Processing Systems 36, 2022. (NeurIPS 2022) (Spotlight, 5% acceptance rate)

Fast Composite Optimization and Statistical Recovery in Federated Learning
Yajie Bao^{#}, Michael Crawshaw^{#}, Shan Luo, Mingrui Liu.
Proceedings of 39th International Conference on Machine Learning, 2022. (ICML 2022)

On the Initialization for ConvexConcave Minmax Problems
Mingrui Liu, Francesco Orabona.
Algorithmic Learning Theory, 2022. (ALT 2022)

On the Last Iterate Convergence of Momentum Methods
Xiaoyu Li, Mingrui Liu, Francesco Orabona.
Algorithmic Learning Theory, 2022. (ALT 2022)

Generalization Guarantee of SGD for Pairwise Learning
Yunwen Lei, Mingrui Liu, Yiming Ying.
Advances in Neural Information Processing Systems 35, 2021. (NeurIPS 2021)
Last update: 04282023