I know there is lots of difference between .eval() and .train() in Dropout layer and BatchNorm layer.
So I write two simple method to locate the reason.
Following is my experiment.
net.eval(),test ,f1_score=0.77 ,It is same with the above result.
net.train()，test，f1_score=77%（the result is different each time as there is randomness in dropout layer）
net.eval(),test,f1_score=0.70 .This time the result is degradation largely.
I don’t know why this result is differ from the above result 0.77???
Is there other difference between .train() and .eval()?
or it is a bug of pytorch???