torch.Tensor.backward — PyTorch 2.7 documentation (original) (raw)

Tensor.backward(gradient=None, retain_graph=None, create_graph=False, inputs=None)[source][source]

Computes the gradient of current tensor wrt graph leaves.

The graph is differentiated using the chain rule. If the tensor is non-scalar (i.e. its data has more than one element) and requires gradient, the function additionally requires specifying a gradient. It should be a tensor of matching type and shape, that represents the gradient of the differentiated function w.r.t. self.

This function accumulates gradients in the leaves - you might need to zero.grad attributes or set them to None before calling it. See Default gradient layoutsfor details on the memory layout of accumulated gradients.

Parameters