Hi everyone.
How do I implement backward view algorithm of TD(lambda) of function approximate.
what is the loss function? or we have to update weight manually by the formula?
are there some elegant method to update grad?
Hi everyone.
How do I implement backward view algorithm of TD(lambda) of function approximate.
what is the loss function? or we have to update weight manually by the formula?
are there some elegant method to update grad?