Dropout regularization
Set a probability
for each node that will be eliminated to reduce weight reliance on a particular node
With the inverted dropout technique, at test time:
You do not apply dropout (do not randomly eliminate units) and do not keep the 1/keep_prob factor in the calculations used in training
Increasing the parameter keep_prob from (say) 0.5 to 0.6 will likely cause the following:
Reducing the regularization effect
Causing the neural network to end up with a lower training set error (overfitting)
Last updated