Invited Talk: A Phase Diagram for Deep Learning Unifying Jamming, Feature Learning and Lazy Training

Speaker: Matthieu Wyart, EPFL
Talk title: A Phase Diagram for Deep Learning Unifying Jamming, Feature Learning and Lazy Training

Time: Wednesday, April 21, 8:00am-9:00am (PT)

Abstract:
Deep learning algorithms are responsible for a technological revolution in a variety of tasks, yet understanding why they work remains a challenge.  Puzzles include that (i) learning corresponds to minimizing a loss in high dimension, which is in general not convex and could well get stuck in bad minima. (ii) Deep learning predicting power increases with the number of fitting parameters, even in a regime where data are perfectly fitted. I will review recent results on these questions based on analogies with physical systems and  scaling arguments testable on real data. For classification, the landscape in deep learning displays a sharp “jamming” transition and becomes glassy as the number of parameters is lowered. This transition also occurs in the packing problem of non-spherical particles.
In the over-parametrized regime  where the landscape has many flat directions,  learning can operate in two regimes “Feature Learning” and “Lazy training” depending on the scale of initialisation. I will provide and test a quantitative explanation as to why performance increases  with the number of parameters in both regimes. I will discuss the relative merits of these regimes based on empirical evidence and simple models.

 

Return to workshop schedule