In the following,

```
x_6 = torch.cat((x_1, x_2_1, x_3_1, x_5_1), dim=-3)
Sizes of tensors x_1, x_2_1, x_3_1, x_5_1 are
torch.Size([1, 256, 7, 7])
torch.Size([1, 256, 7, 7])
torch.Size([1, 256, 7, 7])
torch.Size([1, 256, 7, 7]) respectively.
The size of x_6 turns out to be torch.Size([1, 1024, 7, 7])
```

I couldn’t understand this concatenation along a negative dimension(-3 in this case).

What exactly is happening here?

How does the same go if dim = 3?

Is there any constraint on dim for a given set of tensors?