Is there a memory profiler out there that can output the memory consumed by GPU at every line of the model training and also output the memory consumed by each tensor in the GPU?
Is there a memory profiler out there that can output the memory consumed by GPU at every line of the model training and also output the memory consumed by each tensor in the GPU?