I’m pretty new to Pytorch. I have two questions, hope you can help me.
Imagine the following sequence:
input = [[1,2,3,4,5,6,7,8,9,10]]
Let’s say I want to learn this sequence using 2 look backs. So I would have:
input = [[1,2], [2,3], [3,4]....] and so on.
So my first question is, how do I have pass the input to a LSTM? I know I could shape the input like
(seq_len, batch_size, features) so here it would be
(2, 9, 1), right?
Does the following has the same meaning? So i loop over the batches and pass them to the LSTM but with
batch_size=1. After all batches I do the
... for i in range(input.size(0)): x = [input[i].view(2,1,1) seq.forward(x) .... .... loss.backward()
Is there a difference of setting the
batch_size parameter or just looping over the batches?
- Let’s say I want to predict a value out of 4 look backs. And I have 200 samples of these sequences. How do I have to build the input vector now? Does the number of training samples acutally stands for the batch_size?