PhD Electrical and Computer Engineering
Purdue University
MS Electrical and Computer Engineering
ShanghaiTech
BS Electrical Engineering
Shanghai University
My work focuses on LLM post-training and Multi-agent LLM systems:
PhD, Purdue University — Electrical & Computer Engineering (Aug 2023 — Present)
Working on LLM post-training and reasoning, with an emphasis on RL-based alignment, agent collaboration, and communication-efficient fine-tuning for distributed on-device LLM systems.
Research Intern, Optimization for Machine Learning Lab, KAUST (Aug 2022 — Feb 2023)
Working on optimization theory and its applications on machine learning, especially primal-dual algorithm.
MS, ShanghaiTech University — Electrical & Computer Engineering (2020 — 2023)
Working on optimization theory and its applications on communication system.

Continual post-training framework (DA-GRPO) that jointly learns on-device SLM policies and budget-aware cloud offloading decisions to reduce forgetting while respecting assistance constraints.

RL-based framework that enables on-device LLMs to decide when to invoke cloud models, jointly learning routing and post-training to balance accuracy and compute.

Personalization algorithm for multi-task, multi-modal foundation models that combines architecture-aware replacement with post-FL knowledge distillation.

Communication-efficient framework that uses sketching to adapt LoRA ranks per device, enabling collaborative on-device LLM fine-tuning under heterogeneous resources.

Multi-timescale gradient correction method that stabilizes hierarchical FL across client and group levels, with convergence guarantees robust to data heterogeneity.

HIST framework that partitions large models into submodels across cells to reduce computation, storage, and communication while meeting convergence guarantees.

FedZO algorithm for derivative-free federated optimization, with AirComp-assisted variant to preserve convergence under noisy wireless aggregation.