Tu Vu
I am an Assistant Professor at Virginia Tech (VT) and a Faculty Researcher at Google. At VT, I am also affiliated with the Sanghani Center for Artificial Intelligence & Data Analytics. Prior to joining VT, I held the position of Research Scientist at Google DeepMind for a year after receiving my PhD in Computer Science from the University of Massachusetts Amherst, advised by Mohit Iyyer.
🔍 My research aims to develop effective and efficient methods for advancing and democratizing artificial intelligence in the era of large language models (LLMs). Specific areas of focus include:
- Deep thinking: Developing models that explore diverse reasoning paths and synthesize novel, creative, and high-quality solutions to complex problems
- Agentic memory and context engineering: Developing mechanisms that allow agents to store, retrieve, and use information efficiently over long contexts
- Efficient transfer and adaptation: Reusing learned knowledge across tasks, languages, modalities, or models to adapt effectively in new or low-resource settings with minimal computational cost, data, and storage
- Efficient model updating: Developing methods that keep models up-to-date and responsive to new or evolving information while reasoning effectively over conflicting or manipulated retrieved inputs
⭐ For prospective PhD students
I plan to recruit one new PhD student every year. If you are interested in joining my group, please apply to the VT Graduate School and list me as a potential advisor. Please also check out the application deadlines and information for prospective students. Due to the high volume of emails I receive, I may not be able to respond to each one individually; please don't be discouraged — I may still review your application.
⭐ For Undergraduate and Masters students at VT
I am happy to collaborate on research with current VT students who have at least one full academic year until graduation. If you are interested, feel free to email me. I will follow up if there is a good fit.
Recent news
| Oct. 2025 | |
|---|---|
| Oct. 2025 | |
| Oct. 2025 | |
| Aug. 2025 | |
| Aug. 2025 | |
| Aug. 2025 | |
| Jun. 2025 | |
| Jun. 2025 | |
| Jun. 2025 | |
| Apr. 2025 | |
| Mar. 2025 | |
| Nov. 2024 | |
| Nov. 2024 | ✈️ Attended EMNLP 2024 in Miami, Florida 🌴 |
| Nov. 2024 | |
| Oct. 2024 | |
| Oct. 2024 | |
| Sep. 2024 | |
| Aug. 2024 | |
| Jul. 2024 | |
| May. 2024 | |
| Feb. 2024 | |
| Jan. 2024 | |
| Nov. 2023 | |
| Oct. 2023 | |
| Aug. 2023 | |
| Jul. 2023 | |
Teaching
Advisees
Group:| Rituraj Sharma (1st year MS student) | |
| Noah Provenzano (2nd year MS student) | |
| Weiyuan Chen (1st year PhD student) | |
| Rishab Balasubramanian (2nd year PhD student) | |
| Thinh Pham (2nd year PhD student) | |
| Yu-Min Tseng (1st year PhD student) | |
| Quyet Do (2nd year PhD student) | |
| Pin-Jie Lin (2nd year PhD student) | |
| Jing Chen (1st year PhD student) | |
| Nguyen Nguyen (Junior student) |
| Zhenting Qi (Student Researcher @ Google, Summer - Fall 2025 → PhD student @ Harvard) | |
| Prateek Yadav (Research Intern @ Google DeepMind, Summer 2024 — Spring 2025 → Research scientist @ Meta Superintelligence Labs) | |
| Simeng Han (Student Researcher @ Google DeepMind, Summer 2024 — Spring 2025 → Postdoc @ Stanford) | |
| Salaheddin Alzubi (Masters student @ UMass Amherst, Fall 2022 — Spring 2023 → Research scientist @ Sentient Labs) | |
| Dheeraj Mekala (PhD student @ UCSD, Spring — Summer 2022 → Research scientist @ Meta Superintelligence Labs) |
Preprints
Selected publications
For an up-to-date list of my research papers, please see my Google Scholar profile. * denotes equal contribution.- EMNLPFoundational Autoraters: Taming Large Language Models for Better Automatic EvaluationIn Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024// The top-performing generative model on RewardBench as of July 15, 2024, trained only on publicly available data
- Technical reportGemini: A Family of Highly Capable Multimodal ModelsIn arXiv preprint arXiv:2312.11805, 2023// Google AI Blog
- ACLFreshLLMs: Refreshing large language models with search engine augmentationIn Findings of the Association for Computational Linguistics: ACL 2024, 2024// Our dataset and method have inspired or been used for the development of Google’s Gemini, Perplexity.AI’s Online LLMs, You.com, and Contextual AI’s RAG 2.0