About 44,700 results
Open links in new tab
  1. A Gentle Introduction to torch.autograd — PyTorch Tutorials …

    torch.autograd is PyTorch’s automatic differentiation engine that powers neural network training. In this section, you will get a conceptual understanding of how autograd helps a neural …

  2. Automatic differentiation package - torch.autograd — PyTorch 2.9 ...

    Dec 23, 2016 · Autograd’s aggressive buffer freeing and reuse makes it very efficient and there are very few occasions when in-place operations actually lower memory usage by any …

  3. The Fundamentals of Autograd - PyTorch

    PyTorch’s Autograd feature is part of what make PyTorch flexible and fast for building machine learning projects. It allows for the rapid and easy computation of multiple partial derivatives …

  4. Automatic Differentiation with torch.autograd — PyTorch Tutorials …

    To compute those gradients, PyTorch has a built-in differentiation engine called torch.autograd. It supports automatic computation of gradient for any computational graph.

  5. Autograd mechanics — PyTorch 2.9 documentation

    Jan 16, 2017 · Autograd is a reverse automatic differentiation system. Conceptually, autograd records a graph recording all of the operations that created the data as you execute …

  6. Overview of PyTorch Autograd Engine

    Jun 8, 2021 · Formally, what we are doing here, and PyTorch autograd engine also does, is computing a Jacobian-vector product (Jvp) to calculate the gradients of the model parameters, …

  7. Autograd in C++ Frontend - PyTorch

    Most of the autograd APIs in PyTorch Python frontend are also available in C++ frontend, allowing easy translation of autograd code from Python to C++. In this tutorial explore several examples …

  8. torch.autograd.grad — PyTorch 2.9 documentation

    We use the vmap prototype feature as the backend to vectorize calls to the autograd engine so that this computation can be performed in a single call. This should lead to performance …

  9. Compiled Autograd: Capturing a larger backward graph for

    Compiled Autograd is a torch.compile extension introduced in PyTorch 2.4 that allows the capture of a larger backward graph. While torch.compile does capture the backward graph, it does so …

  10. PyTorch: Defining New autograd Functions

    This implementation computes the forward pass using operations on PyTorch Tensors, and uses PyTorch autograd to compute gradients. In this implementation we implement our own custom …