Hello!

I really cannot explain myself what is going on here: I have two Tensors of size (1000, 2, 2) and (1000, 2 , 5) which share the same values for the first two slices of the third dimension, i.e. [:,:,:2].

```
z0 = torch.rand(1000, 2, 2)
z1 = torch.cat([z0, torch.rand(1000, 2, 3)], dim=2)
s0 = z0[:, :, :2].sum()
s1 = z1[:, :, :2].sum()
```

But when I sum over the same slice I do not get the same result, I have a feeling there’s some floating point precision issue but I don’t see how it can apply in this case since the elements of the tensor are exactly equal. Numpy, on the other hand, does not show this behaviour.

```
>>> torch.all(z0[:, :, :2] == z1[:, :, :2])
True
>>> s0 == s1
False
>>> (s0.sum() - s1.sum()).abs().sum().numpy()
0.017578125
>>> z0.numpy()[:, :, :2].sum() == z1.numpy()[:, :, :2].sum()
True
```

Any clue on why is this happening?

EDIT: if I switch the dimensions such that now I am summing over the first, i.e. z1.shape = [5,2,1000] and s1=z1[:2,:,:].sum() (same applies for z0 and s1) I don’t have the numerical issue. Why?