Jump to content

PyTorch

From Wikipedia, the free encyclopedia
(Redirected from Pytorch)

PyTorch
Original authors
  • Adam Paszke
  • Sam Gross
  • Soumith Chintala
  • Gregory Chanan
DeveloperMeta AI
Initial releaseSeptember 2016; 9 years ago (2016-09)[1]
Stable release
2.9.1[2] Edit this on Wikidata / 12 November 2025; 1 day ago (12 November 2025)
Repositorygithub.com/pytorch/pytorch
Written in
Operating system
PlatformIA-32, x86-64, ARM64
Available inEnglish
TypeLibrary for deep learning
LicenseBSD-3[3]
Websitepytorch.org

PyTorch is an open-source deep learning library, originally developed by Meta Platforms and currently developed with support from the Linux Foundation. The successor to Torch, PyTorch provides a high-level API that builds upon optimised, low-level implementations of deep learning algorithms and architectures, such as the Transformer, or SGD. Notably, this API simplifies model training and inference to a few lines of code. PyTorch allows for automatic parallelization of training and, internally, implements CUDA bindings that speed up training further by leveraging GPU resources.

PyTorch utilises the tensor as a fundamental datatype, similarly to NumPy. Training is facilitated by a reversed automatic differentiation system, Autograd, that constructs a directed acyclic graph of the operations (and their arguments) executed by a model during its forward pass. With a loss, Backpropagation is then undertaken.[4]

As of 2025, PyTorch remains one of the most popular deep learning libraries, alongside others such as TensorFlow and Keras.[5] A number of commercial deep learning architectures are built on top of PyTorch, including Tesla Autopilot,[6] Uber's Pyro,[7] Hugging Face's Transformers,[8][9] and Catalyst.[10][11]

History

[edit]

In 2001, Torch was written and released under a GPL license. It was a machine-learning library written in C++ and CUDA, supporting methods including neural networks, support vector machines (SVM), hidden Markov models, etc.[12][13][14] It was improved to Torch7 in 2012.[15] Development on Torch ceased in 2018 and was subsumed by the PyTorch project.[16]

Meta (formerly known as Facebook) operates both PyTorch and Convolutional Architecture for Fast Feature Embedding (Caffe2), but models defined by the two frameworks were mutually incompatible. The Open Neural Network Exchange (ONNX) project was created by Meta and Microsoft in September 2017 for converting models between frameworks. Caffe2 was merged into PyTorch at the end of March 2018.[17] In September 2022, Meta announced that PyTorch would be governed by the independent PyTorch Foundation, a newly created subsidiary of the Linux Foundation.[18]

PyTorch 2.0 was released on 15 March 2023, introducing TorchDynamo, a Python-level compiler that makes code run up to 2x faster, along with significant improvements in training and inference performance across major cloud platforms.[19][20]

PyTorch tensors

[edit]

PyTorch defines a class called Tensor (torch.Tensor) to store and operate on homogeneous multidimensional rectangular arrays of numbers. PyTorch Tensors are similar to NumPy Arrays, but can also be operated on by a CUDA-capable NVIDIA GPU. PyTorch has also been developing support for other GPU platforms, for example, AMD's ROCm[21] and Apple's Metal Framework.[22]

PyTorch supports various sub-types of Tensors.[23]

Note that the term "tensor" here does not carry the same meaning as tensor in mathematics or physics. The meaning of the word in machine learning is only superficially related to its original meaning as a certain kind of object in linear algebra. Tensors in PyTorch are simply multi-dimensional arrays.

PyTorch neural networks

[edit]

PyTorch defines a module called nn (torch.nn) to describe neural networks and to support training. This module offers a comprehensive collection of building blocks for neural networks, including various layers and activation functions, enabling the construction of complex models. Networks are built by inheriting from the torch.nn module and defining the sequence of operations in the forward() function.

Example

[edit]

The following program shows the low-level functionality of the library with a simple example.

import torch
dtype = torch.float
device = torch.device("cpu")  # Execute all calculations on the CPU
# device = torch.device("cuda:0")  # Executes all calculations on the GPU

# Create a tensor and fill it with random numbers
a = torch.randn(2, 3, device=device, dtype=dtype)
print(a)
# Output: tensor([[-1.1884,  0.8498, -1.7129],
#                  [-0.8816,  0.1944,  0.5847]])

b = torch.randn(2, 3, device=device, dtype=dtype)
print(b)
# Output: tensor([[ 0.7178, -0.8453, -1.3403],
#                  [ 1.3262,  1.1512, -1.7070]])

print(a * b)
# Output: tensor([[-0.8530, -0.7183,  2.58],
#                  [-1.1692,  0.2238, -0.9981]])

print(a.sum()) 
# Output: tensor(-2.1540)

print(a[1, 2])  # Output of the element in the third column of the second row (zero-based)
# Output: tensor(0.5847)

print(a.max())
# Output: tensor(0.8498)

The following code-block defines a neural network with linear layers using the nn module.

from torch import nn # Import the nn sub-module from PyTorch

class NeuralNetwork(nn.Module):  # Neural networks are defined as classes
    def __init__(self):  # Layers and variables are defined in the __init__ method
        super().__init__()  # Must be in every network.
        self.flatten = nn.Flatten()   # Construct a flattening layer.
        self.linear_relu_stack = nn.Sequential(  # Construct a stack of layers.
            nn.Linear(28 * 28, 512),  # Linear Layers have an input and output shape
            nn.ReLU(),  # ReLU is one of many activation functions provided by nn
            nn.Linear(512, 512),
            nn.ReLU(),
            nn.Linear(512, 10), 
        )

    def forward(self, x):  # This function defines the forward pass.
        x = self.flatten(x)
        logits = self.linear_relu_stack(x)
        return logits

See also

[edit]

References

[edit]
  1. ^ Chintala, Soumith (1 September 2016). "PyTorch Alpha-1 release". GitHub. Archived from the original on 29 August 2021. Retrieved 19 August 2020.
  2. ^ "Release 2.9.1". 12 November 2025. Retrieved 13 November 2025.
  3. ^ Claburn, Thomas (12 September 2022). "PyTorch gets lit under The Linux Foundation". The Register. Archived from the original on 18 October 2022. Retrieved 18 October 2022.
  4. ^ "Autograd Mechanics". PyTorch Documentation. Retrieved 13 November 2025.
  5. ^ "Top 30 Open Source Projects". github.com. Retrieved 13 November 2025.
  6. ^ Karpathy, Andrej (6 November 2019). "PyTorch at Tesla - Andrej Karpathy, Tesla". YouTube. Archived from the original on 24 March 2023. Retrieved 2 June 2020.
  7. ^ "Uber AI Labs Open Sources Pyro, a Deep Probabilistic Programming Language". Uber Engineering Blog. 3 November 2017. Archived from the original on 25 December 2017. Retrieved 18 December 2017.
  8. ^ PYTORCH-TRANSFORMERS: PyTorch implementations of popular NLP Transformers, PyTorch Hub, 1 December 2019, archived from the original on 11 June 2023, retrieved 1 December 2019
  9. ^ "Ecosystem Tools". pytorch.org. Archived from the original on 18 July 2023. Retrieved 18 June 2020.
  10. ^ GitHub - catalyst-team/catalyst: Accelerated DL & RL, Catalyst-Team, 5 December 2019, archived from the original on 22 December 2019, retrieved 5 December 2019
  11. ^ "Ecosystem Tools". pytorch.org. Archived from the original on 18 July 2023. Retrieved 4 April 2020.
  12. ^ "Torch Tutorial", Ronan Collobert, IDIAP, 2002-10-02
  13. ^ R. Collobert, S. Bengio and J. Mariéthoz. Torch: a modular machine learning software library. Technical Report IDIAP-RR 02-46, IDIAP, 2002.
  14. ^ "Torch Library". Archived from the original on 31 October 2001.
  15. ^ Collobert, Ronan; Kavukcuoglu, Koray; Farabet, Clément (2012), Montavon, Grégoire; Orr, Geneviève B.; Müller, Klaus-Robert (eds.), "Implementing Neural Networks Efficiently", Neural Networks: Tricks of the Trade: Second Edition, Berlin, Heidelberg: Springer, pp. 537–557, doi:10.1007/978-3-642-35289-8_28, ISBN 978-3-642-35289-8, retrieved 10 June 2025
  16. ^ torch/torch7, Commit fd0ee3b, 2018-07-02
  17. ^ "Caffe2 Merges With PyTorch". 2 April 2018. Archived from the original on 30 March 2019. Retrieved 2 January 2019.
  18. ^ Edwards, Benj (12 September 2022). "Meta spins off PyTorch Foundation to make AI framework vendor neutral". Ars Technica. Archived from the original on 13 September 2022. Retrieved 13 September 2022.
  19. ^ "Dynamo Overview".
  20. ^ "PyTorch 2.0 brings new fire to open-source machine learning". VentureBeat. 15 March 2023. Archived from the original on 16 March 2023. Retrieved 16 March 2023.
  21. ^ "Installing PyTorch for ROCm". rocm.docs.amd.com. 9 February 2024.
  22. ^ "Introducing Accelerated PyTorch Training on Mac". pytorch.org. Archived from the original on 29 January 2024. Retrieved 4 June 2022.
  23. ^ "An Introduction to PyTorch – A Simple yet Powerful Deep Learning Library". analyticsvidhya.com. 22 February 2018. Archived from the original on 22 October 2019. Retrieved 11 June 2018.
[edit]