Tu Vu

prof_pic.jpg

I am a Research Scientist at Google DeepMind and an incoming Assistant Professor at Virginia Tech. Previously, I received my PhD in Computer Science at the University of Massachusetts Amherst, advised by Mohit Iyyer. I also spent three years (2020–2023) as a research intern/student researcher at Google DeepMind and Google Research. My research aims to develop effective and efficient methods for advancing and democratizing artificial intelligence in the era of large language models (LLMs).

My current research focuses on addressing limitations of LLMs and developing novel learning paradigms. Specific areas of focus include:

  • In-context learning and tool-use LLMs: injecting knowledge into LLM prompts and augmenting LLMs with external tools
  • Instruction tuning: enhancing LLMs’ instruction-following capabilities
  • Parameter-efficient transfer learning: efficiently transferring knowledge across tasks, languages, and modalities
  • Long-context modeling: designing efficient model architectures for long sequences
  • Advanced planning and reasoning: improving LLMs’ ability to solve complex reasoning problems
  • Few-shot learning: learning from limited human-labeled data.

If you are interested in doing a PhD at Virginia Tech and joining my lab, please apply to the Virginia Tech Graduate School and list me as a potential advisor. Please also check out the application deadlines and information for prospective students.

Recent news

Feb. 2024 :briefcase: I am now serving as an Area Chair for ACL Rolling Review (ARR)
Jan. 2024 :page_facing_up: Flan-MoE got accepted to ICLR 2024! :tada:
Nov. 2023 :speaking_head: Talk at Graph Neural Networks Reading Group, Google
Oct. 2023 :page_facing_up: New preprint on LLM factuality
Aug. 2023 :briefcase: I joined Google Research in Mountain View, CA as a Research Scientist
Jul. 2023 :mortar_board: I successfully defended my PhD thesis! :tada: :champagne:

Advisees

// Group
Quyet Do (incoming PhD student @ Virginia Tech)
Thinh Pham (incoming PhD student @ Virginia Tech)
Rishab Balasubramanian (incoming PhD student @ Virginia Tech)
Linus Pin-Jie Lin (incoming PhD student @ Virginia Tech)
// Others
Prateek Yadav (Research Intern @ Google Gemini, Summer 2024)
Simeng (Shirley) Han (Student Researcher @ Google DeepMind, Summer 2024)
Dheeraj Mekala (PhD student @ UCSD, Spring & Summer 2022; one paper accepted to EMNLP 2022)

Selected publications

For an up-to-date list of my research papers, please see my Google Scholar profile or my Semantic Scholar profile.
  1. Preprint
    FreshLLMs: Refreshing large language models with search engine augmentation
    Tu VuMohit IyyerXuezhi WangNoah ConstantJerry WeiJason WeiChris TarYun-Hsuan SungDenny ZhouQuoc Leand Thang Luong
    arXiv preprint arXiv:2310.03214, 2023
    // Our dataset and method have inspired or been used for the development of Google’s Gemini, Perplexity.AI’s Online LLMs, You.com, and Contextual AI’s RAG 2.0
  2. ICML
    The Flan Collection: Designing Data and Methods for Effective Instruction Tuning
    Shayne LongpreLe HouTu VuAlbert WebsonHyung Won ChungYi TayDenny ZhouQuoc V LeBarret ZophJason Weiand Adam Roberts
    In Proceedings of the 40th International Conference on Machine Learning, 2023
    // Google Research Blog
  3. ICLR
    Mixture-of-experts meets instruction tuning: A winning combination for large language models
    Sheng ShenLe HouYanqi ZhouNan DuShayne LongpreJason WeiHyung Won ChungBarret ZophWilliam FedusXinyun ChenTu VuYuexin WuWuyang ChenAlbert WebsonYunxuan LiVincent ZhaoHongkun YuKurt KeutzerTrevor Darrelland Denny Zhou
    Proceedings of the 12th International Conference on Learning Representations, 2024
  4. ACL
    SPoT: Better Frozen Model Adaptation through Soft Prompt Transfer
    Tu VuBrian LesterNoah ConstantRami Al-Rfouand Daniel Cer
    In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022
    // Headlines of Google AI’s Natural Language Accelerated Newsletter Q1, 2022
  5. EMNLP
    Overcoming Catastrophic Forgetting in Zero-Shot Cross-Lingual Generation
    Tu VuAditya BaruaBrian LesterDaniel CerMohit Iyyerand Noah Constant
    In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, 2022
  6. EMNLP
    STraTA: Self-Training with Task Augmentation for Better Few-shot Learning
    Tu VuMinh-Thang LuongQuoc LeGrady Simonand Mohit Iyyer
    In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021
  7. EMNLP
    Exploring and Predicting Transferability across NLP Tasks
    Tu VuTong WangTsendsuren MunkhdalaiAlessandro SordoniAdam TrischlerAndrew Mattarella-MickeSubhransu Majiand Mohit Iyyer
    In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, 2020