I’ve encountered a very strange behaviour of autograd.
I have to functions, say f(X) and g(X). And I want to compute the gradient of g(f(X)) with respect to X, which is a Variable.
Chunks of code
res = torch.norm(f(X)) res.backward()
res = torch.norm(g(X)) res.backward()
But when I do
res = torch.norm(f(g(X))) res.backward()
my jupyter kernel dies.
Do you know what is wrong?
I’ll try to write a minimal working example of f and g and add it later.