Hello, I’m trying to migrate my GAN model code from Keras (backend TF) into Pytorch.
Currently I’m stuck at this operation, where I combined result from an embedding layer and image input thru element-wise multiplication.
In Keras, the code line would be such like this
label_embedding = Flatten()(Embedding(n_class, embed_dim)(label)) model_input = Multiply()([input_img, label_embedding])
label_embedding give a vector of size embed_dim and
input_img is in shape of (h, w, 1), resulting in
model_input in shape of (h, w, embed_dim)
In Pytorch, however, I can’t simply multiply the embedding result because of difference in dimension
self.embedding = nn.Embedding(n_class, embed_dim) model_input = torch.mul(input_img, self.embedding(y))
How do I reshape the embedding vector to match with image input and able to be operated using