I’m a 3rd year PhD candidate in Computer Science at the University of Washington, and a visiting researcher at Facebook AI Research. I’m advised by Noah Smith and Luke Zettlemoyer. I am supported by the Bloomberg Data Science PhD Fellowship.
These days, I’m excited about developing models that are modular and embarrassingly parallel. Much of my research investigates language variation in large datasets, and how the composition of training data affects the overall behavior of language models. I strongly believe that being careful about our data will lead to stronger and more reliable language technologies.
Check out my publications to learn more.
- Oct 2022: Our new paper, “lo-fi: distributed fine-tuning without communication” is live!
- Oct 2022: Three papers (“Whose Language Counts as High Quality”, “M2D2”, and “Nearest Neighbor Zero-Shot Inference”) accepted to EMNLP 2022!
- Sept 2022: Talk at USC
- Aug 2022: Talk at Mosaic ML, on “Branch-Train-Merge”
- Aug 2022: Our new paper “Branch-Train-Merge” just dropped!
- June 2022: Our new paper “Nearest Neighbor Zero-Shot Inference” is live!
- March 2022: Two papers (“DEMix” and “Time Waits for No One!”) accepted to NAACL 2022!
- March 2022: Talk at IBM Research Zurich.
- April 2022: I’ll be giving a guest lecture in the Data Processing + Values course at UW on our quality filtering paper.
- January 2022: Our new preprint, “Whose Language Counts As High Quality?”, just dropped!