Google Research, Brain Team
1600 Amphitheatre Parkway
Mountain View, CA, 94043
E-Mail: jaehlee at google dot com
Curriculum Vitae: CV
Before joining Google in 2017, my main research focus was on theoretical high-energy physics. I was a postdoctoral researcher in the Department of Physics & Astronomy at University of British Columbia (UBC) in the String Theory Group. Before that, I completed my PhD in Center for Theoretical Physics (CTP) at MIT working on theoretical physics.
[NEW!] June 2022: Our paper Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models is on ArXiv! Benchmark contains 204 tasksa and was contributed by 444 authors across 132 institutions. It has been already used to evluate recent large language models such as [PALM],[Gopher], and [Chinchilla].
Sep 2021: Our paper Dataset Distillation with Infinitely Wide Convolutional Networks is accepted at NeurIPS 2021!
Beyond the Imitation Game: Quantifying and extrapolating the capabilities of language models
BIG-bench collaboration, member of Core Contributors
[arXiv: 2206.04615] [https://github.com/google/BIG-bench]
Dataset Distillation with Infinitely Wide Convolutional Networks
Timothy Nguyen, Roman Novak, Lechao Xiao, Jaehoon Lee
Neural Information Processing Systems (NeurIPS), 2021
[arXiv: 2107.13034] [code / dataset] [Google AI Blog]
Explaining Neural Scaling Laws
Yasaman Bahri*, Ethan Dyer*, Jared Kaplan*, Jaehoon Lee*, Utkarsh Sharma*
Exploring the Uncertainty Properties of Neural Networks’ Implicit Priors in the Infinite-Width Limit
Ben Adlam*, Jaehoon Lee*, Lechao Xiao*, Jeffrey Pennington and Jasper Snoek
International Conference on Learning Representations (ICLR), 2021 [code / dataset]
ICML 2020 Workshop on Uncertainty & Robustness in Deep Learning [arXiv: 2010:07355]
Finite Versus Infinite Neural Networks: an Empirical Study
Jaehoon Lee, Samuel S. Schoenholz, Jeffrey Pennington, Ben Adlam, Lechao Xiao, Roman Novak, Jascha Sohl-Dickstein
Neural Information Processing Systems (NeurIPS), 2020. [spotlight]
Neural Tangents: Fast and Easy Infinite Neural Networks in Python
Roman Novak, Lechao Xiao, Jiri Hron, Jaehoon Lee, Alexander A. Alemi, Jascha Sohl-Dickstein, Samuel S. Schoenholz
International Conference on Learning Representation(ICLR), 2020 [spotlight]
[arXiv: 1912.02803] [code]
Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent
Jaehoon Lee*, Lechao Xiao*, Samuel S. Schoenholz, Yasaman Bahri, Jascha Sohl-Dickstein, Jeffrey Pennington
Neural Information Processing Systems (NeurIPS), 2019.
Special Isssue, Journal of Statistical Mechanics: Theory and Experiment, 2020.
[arXiv: 1902.06720] [code1] [code2] [Wikipedia(Neural tangent kernel)]
Measuring the Effects of Data Parallelism on Neural Network Training
Christopher J. Shallue*, Jaehoon Lee*, Joseph Antognini, Jascha Sohl-Dickstein, Roy Frostig, George E. Dahl
Journal of Machine Learning Research, 2019.
Deep Neural Networks as Gaussian Processes
Jaehoon Lee*, Yasaman Bahri*, Roman Novak, Samuel S. Schoenholz, Jeffrey Pennington, Jascha Sohl-Dickstein
International Conference on Learning Representations (ICLR), 2018.
[arXiv: 1711.00165] [code] [Wikipedia(Neural network Gaussian process)]
- Recent research interests include:
- Interplay between physics and machine learning
- Theoretical aspects of deep neural networks
- Scientific and principled study of deep neural networks and their learning algorithms
- Theoretical physics with focus on high energy physics
- Action Editor for TMLR
- Area Chair for NeurIPS
- Reviewer for ICLR / ICML / NeurIPS / JMLR / Neural Computation / Pattern Recognition Letters / Nature Communications / TPAMI
- Organizer for Aspen Winter Conference on Physics for Machine Learning
- Organizer for ICML Workshop on Theoretical Physics for Deep Learning
- Organizer for Vancouver deep learning study group