Tu Vu

prof_pic.jpg

I am an Assistant Professor at Virginia Tech (VT) and a Faculty Researcher at Google. At VT, I am also affiliated with the Sanghani Center for Artificial Intelligence & Data Analytics. Prior to joining VT, I held the position of Research Scientist at Google DeepMind for a year after receiving my PhD in Computer Science from the University of Massachusetts Amherst, advised by Mohit Iyyer.

🔍 My research aims to develop effective and efficient methods for advancing and democratizing artificial intelligence in the era of large language models (LLMs). Current areas of focus include:

  • “Deep Think”: Exploring diverse reasoning paths and synthesizing creative, high-quality solutions
  • Agentic memory and context engineering: Storing, retrieving, and using information efficiently over long contexts
  • Efficient transfer and adaptation: Reusing learned knowledge across tasks, languages, modalities, or models to adapt effectively (e.g., in low-resource settings)
  • Efficient model updating: Keeping models up to date while reasoning robustly over conflicting or manipulated retrieved inputs.
⭐ For prospective PhD students

I plan to recruit one new PhD student every year. If you are interested in joining my group, please apply to the VT Graduate School and list me as a potential advisor. Please also check out the application deadlines and information for prospective students. Due to the high volume of emails I receive, I may not be able to respond to each one individually; please don't be discouraged — I may still review your application.

⭐ For Undergraduate and Masters students at VT

I am happy to collaborate on research with current VT students who have at least one full academic year until graduation. If you are interested, feel free to email me. I may follow up later if there is a good fit, even if I do not respond immediately.


Recent news

Jan. 2026 :page_facing_up: Paper to appear at ICLR 2026 on SealQA (reasoning over conflicting evidence)! :tada:
Oct. 2025 :speaking_head: Lightning talk at the Amazon-Virginia Tech AI Workshop
Oct. 2025 :chart_with_upwards_trend: Received the Amazon - VT faculty research award :pray:
Oct. 2025 :chart_with_upwards_trend: Received research awards from Google DeepMind and Google Research :pray:
Aug. 2025 :speaking_head: Gave two invited lectures at The New Turing Institute’s GStar program
Aug. 2025 :page_facing_up: Paper to appear at EMNLP 2025 on efficient model development! :tada:
Aug. 2025 :speaking_head: Featured invited speaker at the Open AGI Symposium at UC Berkeley, hosted by Sentient
Jun. 2025 :page_facing_up: Paper to appear at TMLR 2025 on model merging at scale! :tada:
Jun. 2025 :speaking_head: Invited guest lecture at The New Turing Institute
Jun. 2025 :page_facing_up: New preprint on a challenge benchmark for LLM reasoning over conflicting evidence
Apr. 2025 :chart_with_upwards_trend: Received the New Faculty Mentoring Grant from VT :pray:
Mar. 2025 :page_facing_up: New preprint on fine-tuning transfer for efficient model development
Nov. 2024 :chart_with_upwards_trend: Received a research gift award from Adobe :pray:
Nov. 2024 ✈️ Attended EMNLP 2024 in Miami, Florida 🌴
Nov. 2024 :speaking_head: Invited talk at Qualcomm Seminar Series
Oct. 2024 :speaking_head: Invited talk at Mila / McGill NLP seminar
Oct. 2024 :page_facing_up: New preprint on model merging at scale
Sep. 2024 :page_facing_up: Paper to appear at EMNLP 2024 on foundational autoraters (FLAMe)! :tada:
Aug. 2024 :briefcase: Started my professorship at Virginia Tech
Jul. 2024 :page_facing_up: New preprint on Foundational Autoraters (FLAMe)
May. 2024 :page_facing_up: FreshLLMs got accepted to ACL 2024 Findings! :tada:
Feb. 2024 :briefcase: I am now serving as an Area Chair for ACL Rolling Review (ARR)
Jan. 2024 :page_facing_up: Flan-MoE got accepted to ICLR 2024! :tada:
Nov. 2023 :speaking_head: Invited talk at Graph Neural Networks Reading Group, Google
Oct. 2023 :page_facing_up: New preprint on LLM factuality (FreshLLMs)
Aug. 2023 :briefcase: Joined Google DeepMind in Mountain View, CA as a Research Scientist
Jul. 2023 :mortar_board: Successfully defended my PhD thesis! :tada: :champagne:

Teaching


Advisees

Group:
Mikaela Cankosyan (Senior student)
Rituraj Sharma (1st year MS student)
Noah Provenzano (2nd year MS student)
Weiyuan Chen (1st year PhD student)
Rishab Balasubramanian (2nd year PhD student)
Thinh Pham (2nd year PhD student)
Yu-Min Tseng (1st year PhD student)
Quyet Do (2nd year PhD student)
Pin-Jie Lin (2nd year PhD student)
Nguyen Nguyen (Junior student)
Others:
Zhenting Qi (Student Researcher @ Google, Summer - Fall 2025 → PhD student @ Harvard)
Prateek Yadav (Research Intern @ Google DeepMind, Summer 2024 — Spring 2025 → Research scientist @ Meta Superintelligence Labs)
Simeng Han (Student Researcher @ Google DeepMind, Summer 2024 — Spring 2025 → Postdoc @ Stanford)
Salaheddin Alzubi (Masters student @ UMass Amherst, Fall 2022 — Spring 2023 → Research scientist @ Sentient Labs)
Dheeraj Mekala (PhD student @ UCSD, Spring — Summer 2022 → Research scientist @ Meta Superintelligence Labs)

Preprints

    Selected publications

    For an up-to-date list of my research papers, please see my Google Scholar profile. * denotes equal contribution.
    1. ICLR
      SealQA: Raising the Bar for Reasoning in Search-Augmented Language Models
      Thinh PhamNguyen NguyenPratibha ZunjareWeiyuan ChenYu-Min Tsengand Tu Vu
      In Proceedings of the 14th International Conference on Learning Representations, 2026
      // Our benchmark dataset has been used by Google’s Gemini, DeepSeek, and Kimi
    2. EMNLP
      Efficient Model Development through Fine-tuning Transfer
      Pin-Jie LinRishab BalasubramanianFengyuan LiuNikhil Kandpaland Tu Vu
      In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, 2025
      Oral presentation, rating 4.5/5 by area chairs
    3. EMNLP
      Foundational Autoraters: Taming Large Language Models for Better Automatic Evaluation
      Tu Vu*Kalpesh Krishna*Salaheddin AlzubiChris TarManaal Faruquiand Yun-Hsuan Sung
      In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024
      // The top-performing generative model on RewardBench as of July 15, 2024, trained only on publicly available data
    4. TMLR
      What Matters for Model Merging at Scale?
      Prateek YadavTu VuJonathan LaiAlexandra ChronopoulouManaal FaruquiMohit Bansaland Tsendsuren Munkhdalai
      In Transactions on Machine Learning Research, 2025
    5. Technical report
      Gemini: A Family of Highly Capable Multimodal Models
      Google Gemini Team: Rohan AnilSebastian BorgeaudYonghui WuJean-Baptiste AlayracJiahui YuRadu SoricutJohan SchalkwykAndrew DaiAnja Hauthand  others including Tu Vu
      In arXiv preprint arXiv:2312.11805, 2023
      // Google AI Blog
    6. ACL
      FreshLLMs: Refreshing large language models with search engine augmentation
      Tu VuMohit IyyerXuezhi WangNoah ConstantJerry WeiJason WeiChris TarYun-Hsuan SungDenny ZhouQuoc Leand Thang Luong
      In Findings of the Association for Computational Linguistics: ACL 2024, 2024
      // Our dataset and method have inspired or been used for the development of Google’s Gemini, Perplexity.AI’s Online LLMs, You.com, and Contextual AI’s RAG 2.0
    7. ICML
      The Flan Collection: Designing Data and Methods for Effective Instruction Tuning
      Shayne LongpreLe HouTu VuAlbert WebsonHyung Won ChungYi TayDenny ZhouQuoc V LeBarret ZophJason Weiand Adam Roberts
      In Proceedings of the 40th International Conference on Machine Learning, 2023
      // Google Research Blog
    8. ICLR
      Mixture-of-experts meets instruction tuning: A winning combination for large language models
      Sheng ShenLe HouYanqi ZhouNan DuShayne LongpreJason WeiHyung Won ChungBarret ZophWilliam FedusXinyun ChenTu VuYuexin WuWuyang ChenAlbert WebsonYunxuan LiVincent ZhaoHongkun YuKurt KeutzerTrevor Darrelland Denny Zhou
      In Proceedings of the 12th International Conference on Learning Representations, 2024
    9. ACL
      SPoT: Better Frozen Model Adaptation through Soft Prompt Transfer
      Tu VuBrian LesterNoah ConstantRami Al-Rfouand Daniel Cer
      In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022
      // Headlines of Google AI’s Natural Language Accelerated Newsletter Q1, 2022
    10. EMNLP
      Overcoming Catastrophic Forgetting in Zero-Shot Cross-Lingual Generation
      Tu VuAditya BaruaBrian LesterDaniel CerMohit Iyyerand Noah Constant
      In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022
    11. EMNLP
      STraTA: Self-Training with Task Augmentation for Better Few-shot Learning
      Tu VuThang LuongQuoc LeGrady Simonand Mohit Iyyer
      In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021
    12. EMNLP
      Exploring and Predicting Transferability across NLP Tasks
      Tu VuTong WangTsendsuren MunkhdalaiAlessandro SordoniAdam TrischlerAndrew Mattarella-MickeSubhransu Majiand Mohit Iyyer
      In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020