Why the initialization of weights and bias should be chosen around 0 – datascience.stackexchange.com

I read this: To train our neural network, we will initialize each parameter W(l)ijWij(l) and each b(l)ibi(l) to a small random value near zero (say according to a Normal(0,ϵ2)Normal(0,ϵ2) ...

from Hot Questions - Stack Exchange OnStackOverflow
via Blogspot

Share this

Artikel Terkait

0 Comment to "Why the initialization of weights and bias should be chosen around 0 – datascience.stackexchange.com"