I am a PhD student at Princeton University and a Princeton AI Lab Fellow, working with Prof. Mengdi Wang, Prof. Andrew Yao, and Prof. Quanquan Gu, where my research focuses on building scalable and capable large language models (LLMs) and multimodal foundation models. My work explores methods to improve LLM reasoning, data curation and algorithms for foundation models, as well as the development of new attention mechanisms, positional encodings, and model architectures.
I have also been a visiting PhD student at the UCLA AGI Lab, and as a Top Seed researcher with the Seed LLM (Foundation Model) Team, working on LLM and MLLM pretraining and scaling. Previously, I earned a Master’s degree and PhD candidacy in Computer Science from IIIS at Tsinghua University, working with Prof. Andrew Yao, and a Bachelor of Science in Mathematics and Computer Science from Yuanpei College at Peking University.
I am currently exploring opportunities at Frontier AI Labs and would be pleased to discuss potential collaborations via .