Is it correct to use data in the maximum value ?
the inputs are features extracted from VGG and I want to normalize them cross channel between o and 1. knowing that the real features should not have gradient.
it is my own implementation, So I’m not sure weather it is the correct way or not.
the idea in general, I’m extracting VGG features from my generated image and the target image, but these features vary in scale, I want to normalize them between 0 and 1 but in cross channel normalization so each each channel with it is corresponding channel from the target image is normalized by the maximum of these two pair channels.