Tu Vu
 
 I am an Assistant Professor at Virginia Tech (VT) and a Faculty Researcher at Google. At VT, I am also affiliated with the Sanghani Center for Artificial Intelligence & Data Analytics. Prior to joining VT, I held the position of Research Scientist at Google DeepMind for a year after receiving my PhD in Computer Science from the University of Massachusetts Amherst, advised by Mohit Iyyer.
🔍 My research aims to develop effective and efficient methods for advancing and democratizing artificial intelligence in the era of large language models (LLMs). Specific areas of focus include:
- Advancing LLMs: improving LLMs’ critical capabilities, including reasoning and instruction following, and their emergent use in evaluation (e.g., LM-as-a-Judge or LLM-as-a-Critic)
- Efficient model development: reusing learned knowledge or components effectively across settings (e.g., tasks, languages, modalities, or models)
- Search agents / search-augmented LLMs: enabling models to respond accurately to recent events while reasoning effectively over conflicting or manipulated retrieved inputs
- Parameter-efficient adaptation: adjusting LLMs to new distributions (e.g., unseen tasks, domains, or languages) efficiently, especially in low-resource settings.
⭐ For prospective PhD students
I plan to recruit one new PhD student every year. If you are interested in joining my group, please apply to the VT Graduate School and list me as a potential advisor. Please also check out the application deadlines and information for prospective students. Due to the high volume of emails I receive, I may not be able to respond to each one individually; please don't be discouraged — I may still review your application.
⭐ For Undergraduate and Masters students at VT
I am happy to collaborate on research with current VT students who have at least one full academic year until graduation. If you are interested, feel free to email me. I will follow up if there is a good fit.
Recent news
| Aug. 2025 |  One paper to appear at EMNLP 2025 on efficient model development!   | 
|---|---|
| Aug. 2025 |  Featured invited speaker at the Open AGI Symposium at UC Berkeley, hosted by Sentient | 
| Jun. 2025 |  One paper to appear at TMLR 2025 on model merging at scale!   | 
| Jun. 2025 |  Invited guest lecture at The New Turing Institute | 
| Jun. 2025 |  New preprint on a challenge benchmark for LLM reasoning over conflicting evidence | 
| Apr. 2025 |  I received the New Faculty Mentoring Grant from VT   | 
| Mar. 2025 |  New preprint on fine-tuning transfer for efficient model development | 
| Nov. 2024 |  Our lab received a research gift from Adobe   | 
| Nov. 2024 | ✈️ Attended EMNLP 2024 in Miami, Florida 🌴 | 
| Nov. 2024 |  Invited talk at Qualcomm Seminar Series | 
| Oct. 2024 |  Invited talk at Mila / McGill NLP seminar | 
| Oct. 2024 |  New preprint on model merging at scale | 
| Sep. 2024 |  One paper to appear at EMNLP 2024 on foundational autoraters (FLAMe)!   | 
| Aug. 2024 |  I started my professorship at Virginia Tech | 
| Jul. 2024 |  New preprint on Foundational Autoraters (FLAMe) | 
| May. 2024 |  FreshLLMs got accepted to ACL 2024 Findings!   | 
| Feb. 2024 |  I am now serving as an Area Chair for ACL Rolling Review (ARR) | 
| Jan. 2024 |  Flan-MoE got accepted to ICLR 2024!   | 
| Nov. 2023 |  Invited talk at Graph Neural Networks Reading Group, Google | 
| Oct. 2023 |  New preprint on LLM factuality (FreshLLMs) | 
| Aug. 2023 |  I joined Google DeepMind in Mountain View, CA as a Research Scientist | 
| Jul. 2023 |  I successfully defended my PhD thesis!     | 
Teaching
Advisees
Group:| Lewis Bass (2nd year MS student @ VT // latent reasoning) | |
| Weiyuan Chen (1st PhD student @ VT // advanced reasoning) | |
| Jing Chen (1st PhD student @ VT // TBD) | |
| Yu-Min Tseng (1st PhD student @ VT // TBD) | |
| Noah Provenzano (2nd year MS student @ VT // advanced reasoning) | |
| Rituraj Sharma (1st year MS student @ VT // advanced reasoning) | |
| Nguyen Nguyen (Junior student @ VT // LLM search agents) | |
| Thinh Pham (2nd year PhD student @ VT // LLM search agents) | |
| Rishab Balasubramanian (2nd year PhD student @ VT // cross-model knowledge transfer) | |
| Quyet Do (2nd year PhD student @ VT // instruction following) | |
| Pin-Jie Lin (2nd year PhD student @ VT // efficient model development) | 
| Zhenting Qi (Student Researcher @ Google, Summer & Fall 2025) | |
| Prateek Yadav (Research Intern @ Google DeepMind, Summer 2024 — Spring 2025) | |
| Simeng Han (Student Researcher @ Google DeepMind, Summer 2024 — Spring 2025) | |
| Salaheddin Alzubi (Masters student @ UMass Amherst, Fall 2022 — Spring 2023) | |
| Dheeraj Mekala (PhD student @ UCSD, Spring — Summer 2022) | 
Preprints
Selected publications
For an up-to-date list of my research papers, please see my Google Scholar profile. * denotes equal contribution.
-  EMNLPFoundational Autoraters: Taming Large Language Models for Better Automatic EvaluationIn Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024// The top-performing generative model on RewardBench as of July 15, 2024, trained only on publicly available data
-  Technical reportGemini: A Family of Highly Capable Multimodal ModelsIn arXiv preprint arXiv:2312.11805, 2023// Google AI Blog
-  ACLFreshLLMs: Refreshing large language models with search engine augmentationIn Findings of the Association for Computational Linguistics: ACL 2024, 2024// Our dataset and method have inspired or been used for the development of Google’s Gemini, Perplexity.AI’s Online LLMs, You.com, and Contextual AI’s RAG 2.0