Do I need to consider the batch_size when defining own module

You don’t need to consider the batch size when initializing the Modules. The Linear layer for example takes in_features as an argument, which would be dimension 1 for x = torch.randn(10, 20).

However, when you need another view on the Tensor, e.g when you need to flatten the Tensor coming from a Conv2d, you most likely want to keep the batch size and flat all remaining dimensions.
You would do it in the forward method:

x = self.conv(x)
x = x.view(x.size(0), -1) #keep batch size
x = self.fc(x)

Does this explanation make it clearer?

2 Likes