Quoc V. Le
Quoc V. Le | |
|---|---|
| Born | Lê Viết Quốc 1982 (age 43–44) |
| Education | Australian National University Stanford University |
| Known for | seq2seq doc2vec Neural architecture search Google Neural Machine Translation |
| Scientific career | |
| Fields | Machine Learning |
| Institutions | Google Brain Google DeepMind |
| Thesis | Scalable feature learning (2013) |
| Doctoral advisor | Andrew Ng |
| Other academic advisors | Alex Smola |
Lê Viết Quốc (born 1982),[1] or in romanized form Quoc Viet Le, is a Vietnamese-American computer scientist and artificial intelligence researcher. He is a Google Fellow at Google DeepMind and a founding member of the Google Brain project.
Le is best known for his pioneering work in deep learning, particularly in large-scale unsupervised learning, Sequence-to-Sequence (seq2seq[2]) models, and AutoML (neural architecture search). His research laid the foundation for modern machine translation systems, such as Google Translate, and advanced the field of Large Language Models (LLMs).[3][4][5][6]
Education and career
Le was born in Hương Thủy in the Thừa Thiên Huế province of Vietnam.[4] He attended Quốc Học Huế High School[7] before moving to Australia in 2004 to pursue a Bachelor’s degree at the Australian National University. During his undergraduate studies, he worked with Alex Smola on Kernel method in machine learning.[8] In 2007, Le moved to the United States to pursue graduate studies in computer science at Stanford University, where his PhD advisor was Andrew Ng.
In 2011, Le became a founding member of Google Brain along with his then advisor Andrew Ng, Google Fellow Jeff Dean, and researcher Greg Corrado.[4] He led Google Brain’s first major breakthrough: a deep learning algorithm trained on 16,000 CPU cores, which learned to recognize cats by watching YouTube videos—without being explicitly taught the concept of a "cat."[9][10]
In 2014, Le co-proposed two influential models in machine learning. Together with Ilya Sutskever, Oriol Vinyals, he introduced the seq2seq model for machine translation, a foundational technique in natural language processing. In the same year, in collaboration with Tomáš Mikolov, Le developed the doc2vec[11] model for representation learning of documents. Le was also a key contributor of Google Neural Machine Translation system.[12]
In 2017, Le initiated and led the AutoML project at Google Brain, pioneering the use of neural architecture search.[13] This project significantly advanced automated machine learning. This work led to EfficientNet, a family of image recognition models that achieved state-of-the-art accuracy while being significantly smaller and faster than previous models.
In 2020, Le contributed to the development of Meena, later renamed LaMDA, a conversational large language model based on the seq2seq architecture.[14] In 2022, Le and coauthors published chain-of-thought prompting, a method that enhances the reasoning capabilities of large language models.[15]
In 2024, Le contributed to the development of AlphaGeometry, an AI system that solves complex geometry problems at a level approaching a human International Mathematical Olympiad (IMO) gold-medalist. The system, published in Nature, demonstrated the ability to solve 25 out of 30 Olympiad geometry problems, significantly outperforming previous state-of-the-art automated theorem provers.[16]
Honors and awards
Le was named MIT Technology Review's innovators under 35 in 2014.[17] He has been interviewed by and his research has been reported in major media outlets including Wired,[5] the New York Times,[18] the Atlantic,[19] and the MIT Technology Review.[20] His work on the 2012 Google Brain project, which pioneered large-scale unsupervised learning, received an ICML Test of Time Honorable Mention Award in 2022.[21] Le was named an Alumni Laureate of the Australian National University School of Computing in 2022.[22] In 2024, Le received the NeurIPS Test of Time Award for the seminal paper "Sequence to Sequence Learning with Neural Networks" (co-authored with Ilya Sutskever and Oriol Vinyals). The award committee described the work as a "cornerstone" of modern AI, noting that it established the encoder-decoder architecture and laid the necessary foundation for the large language models (LLMs) and foundation models that followed.[23]
See also
References
- ^ "'Quái kiệt' AI Lê Viết Quốc - người đứng sau thuật toán Transformers của ChatGPT". Viettimes - tin tức và phân tích chuyên sâu kinh tế, quốc tế, y tế (in Vietnamese). 2023-02-09. Retrieved 2023-07-03.
- ^ Sutskever, Ilya; Vinyals, Oriol; Le, Quoc V. (2014-12-14). "Sequence to Sequence Learning with Neural Networks". arXiv:1409.3215 [cs.CL].
- ^ Zoph, Barret; Le, Quoc V. (2017-02-15). "Neural Architecture Search with Reinforcement Learning". arXiv:1611.01578 [cs.LG].
- ^ a b c "Le Viet Quoc, a young Vietnamese engineer who holds Google's brain". tipsmake.com. 24 May 2019. Retrieved 2022-11-24.
- ^ a b Hernandez, Daniela. "A Googler's Quest to Teach Machines How to Understand Emotions". Wired. ISSN 1059-1028. Retrieved 2022-11-25.
- ^ Chow, Rony (2021-06-07). "Quoc V. Le: Fast, Furious and Automatic". History of Data Science. Retrieved 2022-11-26.
- ^ "Fulbright scholars Vietnam - Le Viet Quoc".
- ^ "Meet Le Viet Quoc, a Vietnamese talent at Google". Tuoi Tre News. 2019-02-15. Retrieved 2022-11-25.
- ^ Markoff, John (June 25, 2012). "How Many Computers to Identify a Cat? 16,000". The New York Times.
- ^ Ng, Andrew; Dean, Jeff (2012). "Building High-level Features Using Large Scale Unsupervised Learning". arXiv:1112.6209 [cs.LG].
- ^ Le, Quoc V.; Mikolov, Tomas (2014-05-22). "Distributed Representations of Sentences and Documents". arXiv:1405.4053 [cs.CL].
- ^ "A Neural Network for Machine Translation, at Production Scale". Google Research Blog. 2016-09-27. Retrieved 2023-07-02.
- ^ Zoph, Barret; Le, Quoc V. (2017-02-15). "Neural Architecture Search with Reinforcement Learning". arXiv:1611.01578 [cs.LG].
- ^ Adiwardana, Daniel; Luong, Minh-Thang; So, David R.; Hall, Jamie; Fiedel, Noah; Thoppilan, Romal; Yang, Zi; Kulshreshtha, Apoorv; Nemade, Gaurav; Lu, Yifeng; Le, Quoc V. (2020-01-31). "Towards a Human-like Open-Domain Chatbot". arXiv:2001.09977 [cs.CL].
- ^ "Language Models Perform Reasoning via Chain of Thought". Google Research Blog. 2022-05-22. Retrieved 2023-07-02.
- ^ Trinh, Trieu H.; Wu, Yuhuai; Le, Quoc V.; He, He; Luong, Thang (2024). "Solving olympiad geometry without human demonstrations". Nature. 625 (7995): 476–482. Bibcode:2024Natur.625..476T. doi:10.1038/s41586-023-06747-5. PMC 10794143. PMID 38233616.
- ^ "Quoc Le". MIT Technology Review. Retrieved 2022-11-24.
- ^ Lewis-Kraus, Gideon (2016-12-14). "The Great A.I. Awakening". The New York Times. ISSN 0362-4331. Retrieved 2022-11-26.
- ^ Madrigal, Alexis C. (2012-06-26). "The Triumph of Artificial Intelligence! 16,000 Processors Can Identify a Cat in a YouTube Video Sometimes". The Atlantic. Retrieved 2022-11-26.
- ^ "AI's Language Problem". MIT Technology Review. Retrieved 2022-11-26.
- ^ "ICML Award Test of Time Award". icml.cc.
- ^ "Celebrating 50 years of teaching computer science at ANU". ANU College of Engineering, Computing and Cybernetics. 2 May 2022. Retrieved 2025-06-06.
- ^ "Announcing the NeurIPS 2024 Test of Time Paper Awards". NeurIPS Blog. 2024-11-05. Retrieved 2026-02-07.