I want to iterate over the training samples (172 samples) a fixed number of times (e.g. 512)
trainer.run(train_loader, max_epochs=500, epoch_length=512)
and in my Dataset subclass implementation I’ve defined
__len__ to return the number of samples (172 samples).
But epoch_length does not stop loading at 512, but at 688 samples (4x172). I’ve noticed this behavior by printing a counter when
__getitem__ is called.