Skip to yearly menu bar Skip to main content


Poster

Small nonlinearities in activation functions create bad local minima in neural networks

Chulhee Yun · Suvrit Sra · Ali Jadbabaie

Great Hall BC #17

Keywords: [ neural network ] [ loss surface ] [ spurious local minima ] [ optimization landscape ]


Abstract:

We investigate the loss surface of neural networks. We prove that even for one-hidden-layer networks with "slightest" nonlinearity, the empirical risks have spurious local minima in most cases. Our results thus indicate that in general "no spurious local minim" is a property limited to deep linear networks., and insights obtained from linear networks may not be robust. Specifically, for ReLU(-like) networks we constructively prove that for almost all practical datasets there exist infinitely many local minima. We also present a counterexample for more general activations (sigmoid, tanh, arctan, ReLU, etc.), for which there exists a bad local minimum. Our results make the least restrictive assumptions relative to existing results on spurious local optima in neural networks. We complete our discussion by presenting a comprehensive characterization of global optimality for deep linear networks, which unifies other results on this topic.

Live content is unavailable. Log in and register to view live content