The Applied Study of Over-parameterized Neural Networks

Check out my paper!

Background and abstract

My paper is based on On the Power of Over-parametrization in Neural Networks with Quadratic Activation. This paper by Du and Lee performed the theoretical deductions of the property of over-parameterized one-layer networks with quadratic activation to have all local minima as global minima. My paper performs an empirical analysis of their paper and extended the activation functions to ReLU. My conclusions suggests that 'over-parameterized definition' of their paper is weak as many other cases showcase the same property of global convergence. I explored the boundary for such property in one-layer network with quadratic and ReLU activation.

Sample images

1

NN with Quadratic activation's loss (log)

2

NN with ReLU activation's loss


Evaluation by my Professor

Check out the Evaluation