I completed my MS degree at University of Illinois Urbana-Champaign (UIUC), advised by Prof. Hao Peng. I will join University of Washington as a PhD student in 2024 Fall. I work on Large Language Models.
Research Interest: Natural Language Processing, Large Language Model Pretraining, Reinforcement Learning, Machine Learning, Computer Systems, Distributed Computing.
Pandora: Towards General World Model with Natural Language Actions and Video States
preprint
LLM360: Towards Fully Transparent Open-Source LLMs
preprint
SlimPajama-DC: Understanding Data Combinations for LLM Training
preprint
Language Models Meet World Models: Embodied Experiences Enhance Language Models
NeurIPS 2023
On the Learning of Non-autoregressive Transformers
ICML 2022
Don't Take It Literally: An Edit-Invariant Sequence Loss for Text Generation
NAACL 2022