Does torch.compile guarantee that the generated fusion strategy and Triton kernel are the same every time?

  • Will torch.compile generate different computation graphs under different runtime environments?
  • Can the cache in /tmp/torchinductor_{user} affect the generated results?

I don’t think so as different code and optimizations will be generated for different hardware architectures.