Hi, when Training using BCEWithLogitLoss, we know that it internally apply sigmoid for us, and we don’t need any sigmoid again in our model.
But for inference, do we need to add any sigmoid in the last layer of our model? To make the logits become (0,1)?