What’s the point of changing the mode of
nnReLU operation from inplace to out-of-place in PyTorch Neural style transfer tutorial. I dont think that it would make any difference even if the relu is inplace.
It’s mentioned in the document that,
# The in-place version doesn't play very nicely with the ContentLoss and # StyleLoss we insert below. So we replace with out-of-place # ones here.
Why is that exactly true?