PyTorch all_to_all_single will cause memory leak?

For example, I will call dist.all_to_all_single in a for-loop, this will cause memory footprint increasing like:

But if I comment the line which calls all_to_all_single, the memory footprint increaed slowly like below:

So dist.all_to_all_single will cause memory leak?