Hello everyone, - currently I’m training EfficientNet-B0-ns (it’s kinda new SOTA ResNet) model for some classification task, num_classes = 5.
My GPU, which is RTX 3060Ti allows me to choose batch size up to 64 in order to completely saturate GPU VRAM.
Also I have 6 core Ryzen 5 2600x, - and even with num_workers set to 12 it can’t saturate the GPU (I can see how the model is struggling to compute something, because it’s not ready yet, via tqdm).
Is there any way we can prevent CPU bottlenecking the GPU in case of small models?
These no such problem on bigger models, though.