There is the context manager profile
(PyTorch Profiler — PyTorch Tutorials 1.11.0+cu102 documentation) which allows us to monitor cpu and gpu times. I’d like to integrate it into my script such that I can turn it on / off flexibly:
from torch.profiler import profile
use_prof = True
with profile(enabled=use_prof,...):
# ...
Only there is no enabled
option, as opposed to the torch.autograd.profile
where such option exists (Automatic differentiation package - torch.autograd — PyTorch 1.11.0 documentation).
Is it possible / is there a workaround to flag the torch.profiler.profile
somehow, just like torch.autograd.profile
?
Thanks!