Hi,

I’m trying to compute the sub hessian vector product of a function efficiently. By sub hessian, I mean that I’d like to exctract a sub matrix from the regular hessian and do a vector product with it.

I am aware of the standard trick for computing h.v.p. efficiently and I’d like to do something similar here.

I think one solution could be to stop the gradient for some entries of the tensor X. Hence I could get the “sub” hessian vector product by double differentiation.

Is there a way in pytorch to stop gradient for some entries of a tensor ?

Thanks for your help