Mathematics of Deep Learning Seminar: Francis Bach

Date


Title: Towards a quantitative analysis of gradient descent for infinitely wide two-layer neural networks

Abstract: I will consider fully connected two-layer neural networks with a number of hidden neurons that tends to infinity. For these networks (and with additional technical conditions), it is known that gradient descent with infinitesimal steps on the risk will converge to its global optimum. However, current results are mostly qualitative, with no indication on the number of neurons to achieve the infinite-width limit, or the time to reach a global optimum. The main goal of the talk is to present this open problem. (Based on joint work with Lénaïc Chizat).

Advancing Research in Basic Science and MathematicsSubscribe to Flatiron Institute announcements and other foundation updates

privacy consent banner

Privacy preference

We use cookies to provide you with the best online experience. By clicking "Accept All," you help us understand how our site is used and enhance its performance. You can change your choice at any time here. To learn more, please visit our Privacy Policy.