Skip to content Skip to sidebar Skip to footer

Neural Network Isn't Learning For A First Few Epochs On Keras

I'm testing simple networks on Keras with TensorFlow backend and I ran into an issue with using sigmoid activation function The network isn't learning for first 5-10 epochs, and th

Solution 1:

Activation function suggestion:

In practice, the sigmoid non-linearity has recently fallen out of favor and it is rarely ever used. ReLU is the most common choice, if there are a large fraction of “dead” units in network, try Leaky ReLU and tanh. Never use sigmoid.

Reasons for not using the sigmoid:

A very undesirable property of the sigmoid neuron is that when the neuron’s activation saturates at either tail of 0 or 1, the gradient at these regions is almost zero. In addition, Sigmoid outputs are not zero-centered.

Post a Comment for "Neural Network Isn't Learning For A First Few Epochs On Keras"