Quoc V. Le

Quoc V. Le
Born
Lê Viết Quốc

1982 (age 43–44)
EducationAustralian National University
Stanford University
Known forseq2seq
doc2vec
Neural architecture search
Google Neural Machine Translation
Scientific career
FieldsMachine Learning
InstitutionsGoogle Brain
Google DeepMind
Thesis Scalable feature learning  (2013)
Doctoral advisorAndrew Ng
Other academic advisorsAlex Smola

Lê Viết Quốc (born 1982),[1] or in romanized form Quoc Viet Le, is a Vietnamese-American computer scientist and artificial intelligence researcher. He is a Google Fellow at Google DeepMind and a founding member of the Google Brain project.

Le is best known for his pioneering work in deep learning, particularly in large-scale unsupervised learning, Sequence-to-Sequence (seq2seq[2]) models, and AutoML (neural architecture search). His research laid the foundation for modern machine translation systems, such as Google Translate, and advanced the field of Large Language Models (LLMs).[3][4][5][6]

Education and career

Le was born in Hương Thủy in the Thừa Thiên Huế province of Vietnam.[4] He attended Quốc Học Huế High School[7] before moving to Australia in 2004 to pursue a Bachelor’s degree at the Australian National University. During his undergraduate studies, he worked with Alex Smola on Kernel method in machine learning.[8] In 2007, Le moved to the United States to pursue graduate studies in computer science at Stanford University, where his PhD advisor was Andrew Ng.

In 2011, Le became a founding member of Google Brain along with his then advisor Andrew Ng, Google Fellow Jeff Dean, and researcher Greg Corrado.[4] He led Google Brain’s first major breakthrough: a deep learning algorithm trained on 16,000 CPU cores, which learned to recognize cats by watching YouTube videos—without being explicitly taught the concept of a "cat."[9][10]

In 2014, Le co-proposed two influential models in machine learning. Together with Ilya Sutskever, Oriol Vinyals, he introduced the seq2seq model for machine translation, a foundational technique in natural language processing. In the same year, in collaboration with Tomáš Mikolov, Le developed the doc2vec[11] model for representation learning of documents. Le was also a key contributor of Google Neural Machine Translation system.[12]

In 2017, Le initiated and led the AutoML project at Google Brain, pioneering the use of neural architecture search.[13] This project significantly advanced automated machine learning. This work led to EfficientNet, a family of image recognition models that achieved state-of-the-art accuracy while being significantly smaller and faster than previous models.

In 2020, Le contributed to the development of Meena, later renamed LaMDA, a conversational large language model based on the seq2seq architecture.[14] In 2022, Le and coauthors published chain-of-thought prompting, a method that enhances the reasoning capabilities of large language models.[15]

In 2024, Le contributed to the development of AlphaGeometry, an AI system that solves complex geometry problems at a level approaching a human International Mathematical Olympiad (IMO) gold-medalist. The system, published in Nature, demonstrated the ability to solve 25 out of 30 Olympiad geometry problems, significantly outperforming previous state-of-the-art automated theorem provers.[16]

Honors and awards

Le was named MIT Technology Review's innovators under 35 in 2014.[17] He has been interviewed by and his research has been reported in major media outlets including Wired,[5] the New York Times,[18] the Atlantic,[19] and the MIT Technology Review.[20] His work on the 2012 Google Brain project, which pioneered large-scale unsupervised learning, received an ICML Test of Time Honorable Mention Award in 2022.[21] Le was named an Alumni Laureate of the Australian National University School of Computing in 2022.[22] In 2024, Le received the NeurIPS Test of Time Award for the seminal paper "Sequence to Sequence Learning with Neural Networks" (co-authored with Ilya Sutskever and Oriol Vinyals). The award committee described the work as a "cornerstone" of modern AI, noting that it established the encoder-decoder architecture and laid the necessary foundation for the large language models (LLMs) and foundation models that followed.[23]

See also

References

  1. ^ "'Quái kiệt' AI Lê Viết Quốc - người đứng sau thuật toán Transformers của ChatGPT". Viettimes - tin tức và phân tích chuyên sâu kinh tế, quốc tế, y tế (in Vietnamese). 2023-02-09. Retrieved 2023-07-03.
  2. ^ Sutskever, Ilya; Vinyals, Oriol; Le, Quoc V. (2014-12-14). "Sequence to Sequence Learning with Neural Networks". arXiv:1409.3215 [cs.CL].
  3. ^ Zoph, Barret; Le, Quoc V. (2017-02-15). "Neural Architecture Search with Reinforcement Learning". arXiv:1611.01578 [cs.LG].
  4. ^ a b c "Le Viet Quoc, a young Vietnamese engineer who holds Google's brain". tipsmake.com. 24 May 2019. Retrieved 2022-11-24.
  5. ^ a b Hernandez, Daniela. "A Googler's Quest to Teach Machines How to Understand Emotions". Wired. ISSN 1059-1028. Retrieved 2022-11-25.
  6. ^ Chow, Rony (2021-06-07). "Quoc V. Le: Fast, Furious and Automatic". History of Data Science. Retrieved 2022-11-26.
  7. ^ "Fulbright scholars Vietnam - Le Viet Quoc".
  8. ^ "Meet Le Viet Quoc, a Vietnamese talent at Google". Tuoi Tre News. 2019-02-15. Retrieved 2022-11-25.
  9. ^ Markoff, John (June 25, 2012). "How Many Computers to Identify a Cat? 16,000". The New York Times.
  10. ^ Ng, Andrew; Dean, Jeff (2012). "Building High-level Features Using Large Scale Unsupervised Learning". arXiv:1112.6209 [cs.LG].
  11. ^ Le, Quoc V.; Mikolov, Tomas (2014-05-22). "Distributed Representations of Sentences and Documents". arXiv:1405.4053 [cs.CL].
  12. ^ "A Neural Network for Machine Translation, at Production Scale". Google Research Blog. 2016-09-27. Retrieved 2023-07-02.
  13. ^ Zoph, Barret; Le, Quoc V. (2017-02-15). "Neural Architecture Search with Reinforcement Learning". arXiv:1611.01578 [cs.LG].
  14. ^ Adiwardana, Daniel; Luong, Minh-Thang; So, David R.; Hall, Jamie; Fiedel, Noah; Thoppilan, Romal; Yang, Zi; Kulshreshtha, Apoorv; Nemade, Gaurav; Lu, Yifeng; Le, Quoc V. (2020-01-31). "Towards a Human-like Open-Domain Chatbot". arXiv:2001.09977 [cs.CL].
  15. ^ "Language Models Perform Reasoning via Chain of Thought". Google Research Blog. 2022-05-22. Retrieved 2023-07-02.
  16. ^ Trinh, Trieu H.; Wu, Yuhuai; Le, Quoc V.; He, He; Luong, Thang (2024). "Solving olympiad geometry without human demonstrations". Nature. 625 (7995): 476–482. Bibcode:2024Natur.625..476T. doi:10.1038/s41586-023-06747-5. PMC 10794143. PMID 38233616.
  17. ^ "Quoc Le". MIT Technology Review. Retrieved 2022-11-24.
  18. ^ Lewis-Kraus, Gideon (2016-12-14). "The Great A.I. Awakening". The New York Times. ISSN 0362-4331. Retrieved 2022-11-26.
  19. ^ Madrigal, Alexis C. (2012-06-26). "The Triumph of Artificial Intelligence! 16,000 Processors Can Identify a Cat in a YouTube Video Sometimes". The Atlantic. Retrieved 2022-11-26.
  20. ^ "AI's Language Problem". MIT Technology Review. Retrieved 2022-11-26.
  21. ^ "ICML Award Test of Time Award". icml.cc.
  22. ^ "Celebrating 50 years of teaching computer science at ANU". ANU College of Engineering, Computing and Cybernetics. 2 May 2022. Retrieved 2025-06-06.
  23. ^ "Announcing the NeurIPS 2024 Test of Time Paper Awards". NeurIPS Blog. 2024-11-05. Retrieved 2026-02-07.