mygrad.operation_base.Operation.backward#
- Operation.backward(grad: ndarray, **kwargs)[source]#
Back-propagates the gradient through all of the operation’s inputs, which are stored in the tuple self.variables.
Constant tensors (tensor.constant is True) skipped by this process.
- Parameters:
- gradnumpy.ndarray
The back-propagated total derivative with respect to the present operation (f): d(out)/df