Dropout is scaled by 1/p to keep the expected inputs equal during training and testing.
Have a look at this post for more information.
Where did you notice the 1/(1-p) scaling?
Dropout is scaled by 1/p to keep the expected inputs equal during training and testing.
Have a look at this post for more information.
Where did you notice the 1/(1-p) scaling?