I am a PhD student at Princeton University and a Princeton AI Lab Fellow, working with Prof. Mengdi Wang, Prof. Andrew Yao, and Prof. Quanquan Gu, where my research focuses on building scalable and capable large language models (LLMs) and multimodal foundation models. My work explores how to improve LLM reasoning, develop new attention mechanisms, position encodings, and model architectures, and align their behavior with human preferences through general preference models.
I have also been a visiting PhD student at the UCLA AGI Lab and IIIS, Tsinghua University, and as a Top Seed researcher with the Seed LLM (Foundation Model) Team, working on LLM and MLLM pretraining and scaling. Previously, I earned a Master’s degree and PhD candidacy in Computer Science from IIIS at Tsinghua University, working with Prof. Andrew Yao, and a Bachelor of Science in Mathematics and Computer Science from Yuanpei College at Peking University.
I am currently exploring opportunities at Frontier AI Labs and would be pleased to discuss potential collaborations via .