I am a Research Engineer at Google, working on model quality for Gemini's code capabilities in the post-training phase. I design the training recipes, reward architectures, and evaluation frameworks that define how Gemini reasons about, writes, and ships code at scale.

I led the post-training strategy behind Gemini's #1 rank on the WebDev Arena. Core contributor to Gemini 3.0, Gemini 2.5, and Gemini 2.0 — delivering state-of-the-art results across LiveCodeBench Pro, Terminal Bench 2.0, and SWE-bench. My work spans reinforcement learning for code, reward modeling, and long-horizon agentic evaluations.

Previously, I built enterprise LLMs at Capital One (Llama 2, Mixtral, DPO/RLHF), worked with Google DeepMind on multimodal LLMs for document extraction, and published research on multimodal fact-checking (SIGIR — Best Paper Honorable Mention), hate speech detection (EMNLP), and efficient prompt tuning (ACL). I hold an M.S. from Virginia Tech and a B.S. from D.J. Sanghvi College of Engineering.

Highlights

Selected Papers

Experience

Education

Virginia Tech
M.S. Computer Science (Research) · 2021–2023
Thesis: NLP-based Episodic Future Thinking, funded by NIH
D.J. Sanghvi College of Engineering
B.S. Computer Science · 2016–2020

Service & Honors