About Me

I am a Research Scientist on the Alignment team at FAIR, Meta. I work with Dr. Jason Weston on Reasoning, Memory, and Alignment of Large Language Models. In particular, I do research on Self-improving and Co-improving LLMs by building robust Reward Models and scalable RL recipes.

Previously, I obtained a PhD in Computer Science from the University of North Carolina at Chapel Hill, advised by Prof. Mohit Bansal. My PhD was supported by a Google PhD Fellowship and a Rebecca and Munroe Cobey Fellowship. A list of my publications (grouped by topics) can be found here:

Recent News