I have a list and there are many tensor in the list
I want to turn it to just tensor, and I can put it to dataloader
I use for loop and cat the tensor but it is very slow, data size is about 4,800,000
I have a list and there are many tensor in the list
I want to turn it to just tensor, and I can put it to dataloader
I use for loop and cat the tensor but it is very slow, data size is about 4,800,000
If they’re all the same size, then you could torch.unsqueeze
them in dimension 0 and then torch.cat
the results together.
a = []
for i in range(100000):
a.append(torch.rand(1, 100, 100)
b = torch.Tensor(100000, 100, 100)
torch.cat(a, out=b)
I was always doing something like:
a = [torch.FloatTensor([1]).view(1, -1), torch.FloatTensor([2]).view(1, -1)]
torch.stack(a)
Gives me:
(0 ,.,.) =
1
(1 ,.,.) =
2
[torch.FloatTensor of size 2x1x1]
Actually
I have two list
list 1
a = [[tensor 40], [tensor 40], [tensor 40], …] (2400000 tensor in list each tensor size is 40)
b = [[tensor 40], [tensor 40], [tensor 40], …] (2400000 tensor in list each tensor size is 40)
I want to concat a and b to c
c is a tensor and size is torch.Size([4800000, 40])
I use this method to solve my problem
a = torch.stack(a)
b = torch.stack(b)
c = torch.cat((a, b))
Thank you for all your help !
Or you can just torch.stack(a + b)
It works.
Thank you.
This is how it should be done.
a = torch.stack(a) worked for me
To me it’s sort of unintuitive, why wouldn’t using the tensor class work?
torch.tensor([ torch.tensor([i]).repeat(15) for i in range(0,5)])
the list is the same size and it’s really a matrix/tensor already…but somehow only:
torch.stack([ torch.tensor([i]).repeat(15) for i in range(0,5)])
tensor([[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0],
[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],
[2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2],
[3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3],
[4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4]])
worked.
Btw, is this this most efficient way to do it in a vectorized way?