Online learning with recurrent networks one sample at a time

The above no longer works in current versions of PyTorch as far as I’m aware. It won’t allow you to keep the hidden state like that without detaching it from the graph, as it will complain about freeing up the graph or in-place modifications of the tensors. I’m looking for a way around that, which is why I’ve opened this thread here: How to backpropagate a loss through time-series RNN?. Any help with the matter would be much appreciated!

2 Likes