About Me

I am an assistant professor at Department of Computer Science, George Mason University since Fall 2021. Before that I was a postdoc at Rafik B. Hariri Institute at Boston University from 2020-2021, hosted by Francesco Orabona. I received my Ph.D. at Department of Computer Science, The University of Iowa in August 2020, under the advise of Tianbao Yang. Before that I studied at Institute of Natural Sciences and School of Mathematical Sciences at Shanghai Jiao Tong University. I have also spent time working at industrial research labs, such as IBM research AI and Alibaba DAMO Academy. Here is my Google Scholar Citations.

I am looking for self-motivated PhD students (fully-funded) with strong mathematical ablities and (or) programming skills to solve challenging machine learning problems elegantly with mathematical analysis and empirical studies. The main technical skills we need include mathematical optimization, statistical learning theory, algorithms, and deep learning. If you are interested, please drop me an email with your CV and transcript, and apply our PhD program here. Undergrad and graduate student visitors are also welcome. This link provides an overview of our fast-growing GMU CS department.

Research

My research interests are machine learning, optimization, learning theory, and deep learning. My goal is to design provably efficient algorithms for machine learning problems with strong empirical performance. In particular, I work on

  • Mathematical Optimization for Machine Learning: I focus on designing provably efficient optimization algorithms for machine (deep) learning problems, such as training language models, optimizing complex metrics (e.g., AUC maximization, F-measure optimization), hierarchical optimization (e.g., Generative Adversarial Nets, bilevel optimization), etc.

  • Statistical Learning Theory: I work on improving sample complexity and computational complexity for modern machine learning problems.

  • Large-scale Distributed/Federated Learning: I design efficient scalable learning algorithms for distributed intelligence under various constraints (e.g., communication, privacy, etc.)

  • Machine Learning Applications: continual learning, sample selection, parameter-efficient tuning of foundation models.

  • Recent News

    • (Jan 2025) Two papers were accepted by ICLR 2025. Congratulations to my student Michael!
    • (Nov 2024) I will be serving as an Area Chair for ICML 2025.
    • (Sep 2024) Two papers were accepted by NeurIPS 2024. Congratulations to my students Michael, Xiaochuan, and Jie!
    • (Sep 2024) I will be serving as an Area Chair for AISTATS 2025.
    • (Aug 2024) Glad to give an invited talk at Lehigh University.
    • (May 2024) I will be serving as an Area Chair for NeurIPS 2024.
    • (May 2024) Two papers were accepted by ICML 2024. Congratulations to my students!
    • (Feb 2024) Glad to give an invited talk at Virginia Tech CS Seminar Series about our recent work on optimization for deep autoregressive models.
    • (Jan 2024) One paper about bilevel optimization under unbounded smoothness was accepted by ICLR 2024 as an spotlight (5% acceptance rate). Congrats to my students Jie and Xiaochuan!
    • More News

    Recent Selected Publications [Full List]

    Last update: 01-23-2025