Jiawei Zhang

Jiawei Zhang

Assistant professor in CS @ UW-Madison

University of Wisconsin–Madison

Biography

Jiawei Zhang is an assistant professor in the Department of Computer Sciences at University of Wisconsin–Madison. Previously, he was a postdoctoral fellow supported by The MIT Postdoctoral Fellowship For Engineering Excellence in the Laboratory for Information & Decision Systems (LIDS) at MIT, working with Prof. Asuman Ozdaglar and Prof. Saurabh Amin. He obtained the Ph.D. degree in Computer and Information Engineering from the Chinese University of Hong Kong, Shenzhen, with an advisory of Prof. Zhi-Quan (Tom) Luo, and was honored with the Presidential Award for Outstanding Doctoral Students. He obtained the B.Sc. in Mathematics (Hua Loo-Keng Talent Program) from the University of Science and Technology of China.

Interests
  • Efficient, scalable, generalizable, and robust optimization algorithms for large-scale optimization problems
  • Fundamental theories and algorithms for generative models, including LLMs and diffusion models
  • Optimization and decision-making for engineering applications, including energy systems and communication systems
Education
  • Ph.D. in Computer and Information Engineering

    The Chinese University of Hong Kong, Shenzhen (CUHK-Shenzhen)

  • B.Sc. in Mathematics (Hua Loo-Keng Talent Program)

    University of Science and Technology of China (USTC)

Research Topics

  • Efficient, scalable, generalizable, and robust optimization algorithms for large-scale optimization problems, including LLM training, constrained optimization, and bilevel optimization.

  • Fundamental theories and algorithms for generative models, including LLMs and diffusion models, focusing on efficient and reliable training and inference.

  • Optimization and decision-making for engineering applications, including energy systems and communication systems.

Prospective Students

I am looking for students interested in the research areas mentioned above or related topics about optimization, machine learning and their applications to science and engineering starting in 2025. They should have either a strong background in math and good coding skills, or a strong coding ability and adept theoretical insights. Students with backgrounds in mathematics, statistics, computer science, electronic engineering, and other related fields are welcome to apply.

Recent Publications

(2025). Why GRPO Needs Normalization: A Local-Curvature Perspective on Adaptive Gradients. Annual Conference on Neural Information Processing Systems (NeurIPS 2025) ER Workshop.

Cite

(2025). Population-Proportional Preference Learning from Human Feedback: An Axiomatic Approach. arXiv preprint arXiv:2506.05619.

Cite

(2025). MLorc: Momentum Low-rank Compression for Large Language Model Adaptation . arXiv preprint arXiv:2506.01897.

Cite

(2025). On the convergence analysis of muon. arXiv preprint arXiv:2505.23737.

Cite

Preprint Papers

(2025). Population-Proportional Preference Learning from Human Feedback: An Axiomatic Approach. arXiv preprint arXiv:2506.05619.

Cite

(2025). MLorc: Momentum Low-rank Compression for Large Language Model Adaptation . arXiv preprint arXiv:2506.01897.

Cite

(2025). On the convergence analysis of muon. arXiv preprint arXiv:2505.23737.

Cite

(2022). On the Iteration Complexity of Smoothed Proximal ALM for Nonconvex Optimization Problem with Convex Constraints. arXiv preprint arXiv:2207.06304.

Cite

Services

  • Action Editor, Transactions on Machine Learning Research (TMLR).

  • Area Chair, International Conference on Machine Learning (ICML 2026).

Contact