Tu Vu

I am an Assistant Professor at Virginia Tech (VT). I am also a Faculty Researcher at Google. Prior to joining VT, I held the position of Research Scientist at Google DeepMind for a year after receiving my PhD in Computer Science from the University of Massachusetts Amherst, advised by Mohit Iyyer.
🔍 My research aims to develop effective and efficient methods for advancing and democratizing artificial intelligence in the era of large language models (LLMs). Specific areas of focus include:
- Efficient model development: developing techniques to create more powerful LLMs using less computational power and data
- Parameter-efficient transfer learning: efficiently transferring knowledge across tasks, languages, and modalities
- Search engine augmentation: combining LLMs with search engines to improve their factual accuracy and ability to access up-to-date information
- Instruction tuning: enhancing LLMs’ instruction-following capabilities
- Advanced reasoning: improving LLMs’ ability to solve complex reasoning problems
- LLM-as-a-Judge: developing techniques to enhance LLMs for automatic evaluation.
⭐ For prospective PhD students
I plan to recruit one or two PhD students every year. If you are interested in joining my group, please apply to the VT Graduate School and list me as a potential advisor. Please also check out the application deadlines and information for prospective students. Due to the high volume of emails I receive, I may not be able to respond to each one individually; please don't be discouraged — I may still review your application.
⭐ For Undergraduate and Masters students at VT
I am happy to collaborate on research with current VT students, ideally those who have at least one full calendar year remaining until graduation. If you are interested, feel free to email me or visit during my office hours.
Recent news
Nov. 2024 | ✈️ Attended EMNLP 2024 in Miami, Florida 🌴 |
---|---|
Nov. 2024 | ![]() |
Oct. 2024 | ![]() |
Oct. 2024 | ![]() |
Sep. 2024 | ![]() ![]() |
Aug. 2024 | ![]() |
Jul. 2024 | ![]() |
May. 2024 | ![]() ![]() |
Feb. 2024 | ![]() |
Jan. 2024 | ![]() ![]() |
Nov. 2023 | ![]() |
Oct. 2023 | ![]() |
Aug. 2023 | ![]() |
Jul. 2023 | ![]() ![]() ![]() |
Teaching
Advisees
Group:Weiyuan Chen (Incoming PhD student @ VT) | |
Noah Provenzano (1st year MS student @ VT // advanced reasoning) | |
Quyet Do (1st year PhD student @ VT // instruction following) | |
Rituraj Sharma (Senior student & incoming MS student @ VT // advanced reasoning) | |
Rishab Balasubramanian (1st year PhD student @ VT // long-context LLMs) | |
Nguyen Nguyen (Sophomore student @ VT // search-augmented LLMs) | |
Thinh Pham (1st year PhD student @ VT // search-augmented LLMs) | |
Aninditaa Chauhan (1st year MS student @ VT // multimodal agents) | |
Pin-Jie Lin (1st year PhD student @ VT // efficient model development) |
Prateek Yadav (Research Intern @ Google DeepMind, Summer 2024 — Spring 2025) | |
Simeng Han (Student Researcher @ Google DeepMind, Summer 2024 — Spring 2025) | |
Salaheddin Alzubi (Masters student @ UMass Amherst, Fall 2022 — Spring 2023) | |
Dheeraj Mekala (PhD student @ UCSD, Spring — Summer 2022) |
Selected publications
For an up-to-date list of my research papers, please see my Google Scholar profile. * denotes equal contribution.- EMNLPFoundational Autoraters: Taming Large Language Models for Better Automatic EvaluationIn Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024
- PreprintGemini: A Family of Highly Capable Multimodal ModelsIn arXiv preprint arXiv:2312.11805, 2023
- ACLFreshLLMs: Refreshing large language models with search engine augmentationIn Findings of the Association for Computational Linguistics: ACL 2024, 2024