About Me

I am an assistant professor at Department of Computer Science, George Mason University since Fall 2021. Before that I was a postdoc at Rafik B. Hariri Institute at Boston University from 2020-2021, hosted by Francesco Orabona. I received my Ph.D. at Department of Computer Science, The University of Iowa in August 2020, under the advise of Tianbao Yang. Before that I studied at Institute of Natural Sciences and School of Mathematical Sciences at Shanghai Jiao Tong University. I have also spent time working at industrial research labs, such as IBM research AI and Alibaba DAMO Academy. Here is my Google Scholar Citations.

I am looking for self-motivated PhD students (fully-funded) with strong mathematical ablities and (or) programming skills to solve challenging machine learning problems elegantly with mathematical analysis and empirical studies. The main technical skills we need include mathematical optimization, statistical learning theory, algorithms, and deep learning. If you are interested, please drop me an email with your CV and transcript, and apply our PhD program here. Undergrad and graduate student visitors are also welcome. This link provides an overview of our fast-growing GMU CS department.

Research

My research addresses the fundamental problem of efficiency in machine learning, with a focus on the interplay of optimization and learning. I design algorithms with provable performance guarantees that tackle challenges arising in modern learning paradigms, spanning computational, statistical, and scalability dimensions. My research goal is to design provably efficient algorithms for modern machine learning problems with strong empirical performance. My research topics include:

  • Optimization for Machine Learning: I design and analyze optimization algorithms with rigorous efficiency guarantees for modern learning problems, including the training of language models such as Transformers [NeurIPS'22], hierarchical optimization (e.g., learning problems with minimax and bilevel formulation) [ICLR'24 Spotlight, ICML'24, JMLR'21], optimization methods with adaptivity guarantees and its fundamental limits [ICLR'25, ICLR'20], and global convergence of neural network optimization [NeurIPS'23].

  • Algorithmic Generalization and Feature Learning: I study how optimization and learning dynamics shape both generalization performance and the emergence of features in modern models. My work characterizes the sample and computational complexity of problems such as pairwise learning [NeurIPS'21], and develops feature learning theory for neural networks [ICML'24].

  • Distributed and Scalable Learning: I design scalable algorithms for distributed learning that remain efficient under real-world constraints, such as limited communication [NeurIPS'22 Spotlight], decentralization [NeurIPS'20], and node unavailability [NeurIPS'24]. Recently, I focus on moving beyond black-box analyses to develop principled, fine-grained theoretical frameworks for distributed learning problems [ICML'25, ICLR'25].

  • Adaptive Learning: I work on designing algorithms that learn efficiently by selecting informative data and adapting to data distribution shifts. My work includes coreset selection for data-efficient training [NeurIPS'23], continual learning without catastrophic forgetting [NeurIPS'20 Spotlight].

  • Recent News

    • (May 2025) One paper was accepted by ICML 2025. Congratulations to my student Michael and our collaborators!
    • (March/April 2025) Glad to give invited talks about "Beyond Black-Box Analysis: Understanding the Role of Local Updates in Distributed Learning" at UMN and KAUST. Check the video.
    • (Feb 2025) I will be serving as an Area Chair for NeurIPS 2025.
    • (Jan 2025) Two papers were accepted by ICLR 2025. Congratulations to my student Michael and our collaborators!
    • (Nov 2024) I will be serving as an Area Chair for ICML 2025.
    • (Sep 2024) Two papers were accepted by NeurIPS 2024. Congratulations to my students Michael, Xiaochuan, and Jie!
    • (Sep 2024) I will be serving as an Area Chair for AISTATS 2025.
    • (Aug 2024) Glad to give an invited talk at Lehigh University.
    • (May 2024) I will be serving as an Area Chair for NeurIPS 2024.
    • (May 2024) Two papers were accepted by ICML 2024. Congratulations to my students!
    • (Feb 2024) Glad to give an invited talk at Virginia Tech CS Seminar Series about our recent work on optimization for deep autoregressive models.
    • (Jan 2024) One paper about bilevel optimization under unbounded smoothness was accepted by ICLR 2024 as an spotlight (5% acceptance rate). Congrats to my students Jie and Xiaochuan!
    • More News

    Recent Selected Publications [Full List]

    Last update: 08-01-2025