Mathematical Logics Behind The Weights Initialization

Motivation

When I was working on machine learning tasks, I used to initialize the weights in my model by the default settings of the machine learning too, because I “trust” the “intelligent tool” could always provide the best solution to this. This often worked very well, and I did not spend too much time to understand what the tool did and why it did so.

Why the weights initialization is important?

Think of logistic regression. Let us have an extreme case, if weights are badly chosen so that the linear additive output to the activation function is extremely large or small. From the curve of logistic regression, we immediately know that the derivatives of logistic regression when x is very large or small is extremely small.

Logistic Curve

When we are updating the weights in our final output layer or middle hidden layers using backpropagation, the derivative of the weights always contain the item of the derivative of activation function at certain hidden nodes (You may check my blog, Multi-layer Perceptron on Wikipedia or any machine learning textbooks). If this derivative is very small, the learning of the weights would be extremely slow. It should be noted that this might not be totally overcome by making the learning rate bigger, because the big learning rate might be disastrous when some of the weights whose hidden nodes in the right range could be normally updated. Some logistic regression equation have \(\beta\) term, \(g(x) = \frac{1}{1+\exp(-{\beta}x)}\), this term is the same as learning rate (You may see it after calculating the derivative of logistic regression by yourself).

Therefore, setting up the right weights is very important.

Mathematics behind the weights initialization

I often saw a “gold solution” of the weights initialization on the internet, which states that “a good idea to choose initial weights of a neural network randomly from the range \([-\frac{1}{\sqrt{n}}, \frac{1}{\sqrt{n}}]\)”, where n is the number of hidden nodes in the input layer.

Under the assumption “the inputs are normalized to have a mean of 0 and variance (standard deviation^2) of 1”, the sum of the linear addition of all the outputs from the same layer before activation would have mean of 0 and variance of \(\frac{1}{3}\). At this range, the value of the derivative of the activation function would be reasonable. So our learning efficiency would be good.

But how did we get this?

For uncorrelated two ramdonly variables, \(Var(X+Y) = Var(X) + Var(Y)\), and \(Var(XY) = Var(X)Var(Y)\). This is always true if we have more variables (see Variance Property on Wikipedia). The variance of uniform distribution \([-\frac{1}{\sqrt{n}}, \frac{1}{\sqrt{n}}]\) is \(\frac{1}{3n}\) (see Uniform Distribution Property on Wikipedia). So the sum of the variances of all n nodes in the input layer would have a variance of \(\frac{1}{3n}\times{n}=\frac{1}{3}\). This sum also has mean of 0 (I think I do not have to explain this).

Additional idea?

How about initializing all the weights to zero?

Unfortunately, according to the backpropagation derivative update functions, there would be no update for any of the weights in the hidden layers but not the final output layer (You may check my blog, Multi-layer Perceptron on Wikipedia or any machine learning text books). You would only end up with a “single-layer” linear model that performs very bad.

Mathematical Logics Behind The Weights Initialization

https://leimao.github.io/blog/Weights-Initialization/

Author

Lei Mao

Posted on

05-18-2017

Updated on

05-18-2017

Licensed under


Comments