Hi can anybody tell me when should we use average loss for training? I found that in my case when I used loss = mse (output,label) as the loss for learning curve, it would sometimes increase. But if I used average loss instead, the learning curve looks good. I calculate the average loss with the following:
for epoch in range():
for i, batched_images in enumerate(dataloader):
loss = mse(input,label)
running_loss + = loss.item()
num_images += input.size(0)
…
Aveg_loss = running_loss / num_images
Any information would be appreciated:)