Hello oasjd7!
(Or you can just use torch.nn.CrossEntropyLoss
.)
Please see the following thread for an implementation:
You should use LogSoftmax
. You have to pass the output of Softmax
through log()
anyway to calculate the cross entropy, and the
implementation of LogSoftmax
is numerically more stable than (the
mathematically, but not numerically equivalent) log (Softmax)
.
If you don’t naturally have soft target labels (probabilities across the
classes), I don’t see any value in ginning up soft labels by adding
noise to your 0, 1
(one-hot) labels. Just use CrossEntropyLoss
with your hard labels.
(If your hard labels are encoded as 0, 1
-style one-hot labels you will
have to convert them to integer categorical class labels, as those are
what CrossEntropyLoss
requires.)
Best.
K. Frank