Feature Learning in Infinite-Width Neural Networks

Feature Learning in Infinite-Width Neural Networks

Neyman Seminar
Feb 24, 2021, 04:00 PM - 05:00 PM | Zoom id: 97648161149. No passcode. Evans Hall | Happening As Scheduled
Greg Yang, Microsoft Research

As its width tends to infinity, a deep neural network's behavior under gradient descent can become simplified and predictable (e.g. given by the Neural Tangent Kernel (NTK)), if it is parametrized appropriately (e.g. the NTK parametrization). However, we show that the standard and NTK parametrizations of a neural network do not admit infinite-width limits that can learn representations (i.e....