hello
I have been trying to visualize the outputs of a vgg.16 network. But the output seems to be just wrong. As you know the convoltuion doesnt translate the semantic segment of the picture . like for the folwing picture if the head is on the top part of the picture it should be on top of the picture still after the convolution is done . But it doesnt seem to be the case . I used the following code to extract the intermediate layers .
class vgg16(torch.nn.Module):
def __init__(self, pretrained=True):
super(vgg16, self).__init__()
vgg_pretrained_features = tv.vgg16(pretrained=pretrained).features
self.layerss = torch.nn.Sequential()
for x in range(30):
self.layerss.add_module(str(x), vgg_pretrained_features[x])
self.layerss.eval()
def forward(self, x):
output=[]
for i,layer in enumerate( self.layerss):
# print (i)
x=layer(x)
output.append(x)
return output
model=vgg16()
output=model.forward(img)
import matplotlib.pyplot as plt
plt.imshow(output[0][0][0].detach())
here is the original picture and the ouput of the first chanel of the first layer in vgg:
As you can see the face has moved all the way down and the neckless is all the way up and the overal structure of the picture is broken
I cannot reproduce the issue and I guess you might be calling reshape or view on the input image in order to move the channel dimension, which would interleave the image. If that’s the case, use permute instead.
For an input of
and this code snippet:
class vgg16(torch.nn.Module):
def __init__(self):
super(vgg16, self).__init__()
vgg_pretrained_features = models.vgg16().features
self.layerss = torch.nn.Sequential()
for x in range(30):
self.layerss.add_module(str(x), vgg_pretrained_features[x])
self.layerss.eval()
def forward(self, x):
output=[]
for i,layer in enumerate( self.layerss):
# print (i)
x=layer(x)
output.append(x)
return output
img = PIL.Image.open('./drums.png')
x = TF.resize(img, (224, 224))
x = TF.to_tensor(x)[None, :3, :, :]
model = vgg16()
output = model(x)
import matplotlib.pyplot as plt
plt.imshow(output[0][0][0].detach())
thank you very much for your time and you are right i did use reshape . Though i dont know why would reshape change the structure of a tensor torch . but thank you anyway .