I’m trying to increase the weight of an under sampled class in a binary classification problem.
torch.nn.BCELoss has a
weight attribute, however I don’t quite get it as this weight parameter is a constructor parameter and it is not updated depending on the batch of data being computed, therefore it doesn’t achieve what I need.
What is the correct way of simulating a class weight, similar to the way Keras does?
Solved with a custom loss function:
def weighted_binary_cross_entropy(output, target, weights=None):
if weights is not None:
assert len(weights) == 2
loss = weights * (target * torch.log(output)) + \
weights * ((1 - target) * torch.log(1 - output))
loss = target * torch.log(output) + (1 - target) * torch.log(1 - output)
According to the doc here
weight parameter is a tensor of weight for each example in the batch. Thus, it must have the size equal to the batch size. You can set the weight at the beginning of each batch, for example:
criterion = nn.BCELoss() for batch in data: input, label, weight = batch criterion.weight = weight loss = criterion.forward(predct, label) ...
That’s a neat solution as well… I ended up changing my loss function again to NLLLoss, which supports class weights, and it’s probably the easiest native solution. My custom loss was giving me NaNs towards the end of training and I have no idea why!
Did you apply LogSoftmax before computing the loss. NLLLoss takes log probability as input, not the probability.
Yes, I did change the softmax to a log-softmax. The custom loss however, is working with a regular softmax, but I guess it could be related to the lack of an epsilon term to prevent the system to output a “hard one or zero”.
Thank you. This works for me.
Ref to the c code, there is a safe_log function which returns log(1e-12) if the input is 0.
That may be the numerical unstable. Applying clamp may help:
output = torch.clamp(output,min=1e-8,max=1-1e-8)
loss = pos_weight * (target * torch.log(output)) + neg_weight* ((1 - target) * torch.log(1 - output))
yes, that’s the easiest way… But you can simply use the NLLoss, it supports class weights now
there should be a “-” in loss function
sorry I’m wrong, I ignored the torch.neg:sweat:
Hello thanks for ur costum function i use it but i had this eror
The size of tensor a (32) must match the size of tensor b (2) at non-singleton dimension 1
can u help with this problem ?
Just a quick question. When applying BCELoss with weights, do we need to normalize the weights with the batch size? Or raw weights would be fine?
I’m wondering how you used NLLLoss for a binary classification problem?
Thanks,I just don’t know how to use weight to join NNLLoss.
I think ‘output = torch.clamp(output, 1e-9, 1-1e-9)’ to prevent the output from having 0.
can you explain me how this works ?
I am working on classification problem on celebA dataset, in which, for some features there’s a huge imabalance. How can I rectify this issue with your above mentioned code ?
Just replace BCELoss with CrossEntropyLoss, which has a weight per class, it is probably the easiest solution.
criterion = nn.CrossEntropyLoss(weight= torch.tensor([1, 20.4]).to(device))
Is this okay ?