Skip to yearly menu bar Skip to main content


Invited Talk

Learned optimizers: why they're the future, why they’re hard, and what they can do now

Jascha Sohl-Dickstein
2023 Invited Talk

Abstract

Speaker

Jascha Sohl-Dickstein

Jascha Sohl-Dickstein

I am a principal scientist in Google DeepMind, where I lead a research team with interests spanning machine learning, physics, and neuroscience. I'm most (in)famous for [inventing diffusion models](https://arxiv.org/abs/1503.03585). My recent work has focused on theory of [overparameterized neural networks](https://github.com/google/neural-tangents/wiki/Overparameterized-Neural-Networks:-Theory-and-Empirics), meta-training of [learned optimizers](https://arxiv.org/abs/2009.11243), and [understanding the capabilities of large language models](https://github.com/google/BIG-Bench). Before working at Google I was a visiting scholar in [Surya Ganguli's lab](http://ganguli-gang.stanford.edu/) at Stanford University, and an academic resident at [Khan Academy](http://khanacademy.org/).  I earned my PhD in 2012 in the [Redwood Center for Theoretical Neuroscience](http://redwood.berkeley.edu/) at UC Berkeley, in [Bruno Olshausen's](https://redwood.berkeley.edu/bruno/) lab. Prior to my PhD, I [worked on Mars](http://mars.nasa.gov/mer/home/).

Video

Chat is not available.