Huge gap between cuda allocated and reserved memories

Reserved memory contains the allocated and cached memory. So I think they should be roughly the same after calling torch.cuda.empty_cache(). But I too experienced otherwise. So I’m also looking for an answer. (Already asked in this comment.)