My questions might be too dump for advanced users, sorry in advance.
- In the example tutorials like word_language_model or time_sequence_prediction etc. States of lstm/rnn initialized at each epoch:
hidden = model.init_hidden(args.batch_size)
I tried to remove these in my code and it still worked the same. So, when do we actually need to initialize the states of lstm/rnn?
- Let say I want to use different batch sizes in train, validation and test times. I want to use large batch size in training time to speed up the learning and small batch size in the test time as the number of validation samples are small. Is this okay thing to do or should I fix batch size during training and validation? I believed it should be okay but, I got worse results in validation when I changed batch size going from training to validation.