December 02, 2019
Deep learning frameworks have often focused on either usability or speed, but not both. PyTorch is a machine learning library that shows that these two goals are in fact compatible: it provides an imperative and Pythonic programming style that supports code as a model, makes debugging easy and is consistent with other popular scientific computing libraries, while remaining efficient and supporting hardware accelerators such as GPUs. In this paper, we detail the principles that drove the implementation of PyTorch and how they are reflected in its architecture. We emphasize that every aspect of PyTorch is a regular Python program under the full control of its user. We also explain how the careful and pragmatic implementation of the key components of its runtime enables them to work together to achieve compelling performance. We demonstrate the efficiency of individual subsystems, as well as the overall speed of PyTorch on several common benchmarks.
Written by
Soumith Chintala
Adam Lerer
Edward Yang
Francisco Massa
Gregory Chanan
Junjie Bai
Lu Fang
Sam Gross
Zachary DeVito
Zeming Lin
Adam Paszke
Alban Desmaison
Alykhan Tejani
Andreas Köpf
James Bradbury
Luca Antiga
Martin Raison
Natalia Gimelshein
Sasank Chilamkurthy
Trevor Killeen
Publisher
NeurIPS
Research Topics
December 13, 2022
Michael Auli, Alexei Baevski, Arun Babu, Wei-Ning Hsu
December 13, 2022
November 28, 2022
Nicolas Ballas, Bernhard Schölkopf, Chris Pal, Francesco Locatello, Li Erran, Martin Weiss, Nasim Rahaman, Yoshua Bengio
November 28, 2022
November 27, 2022
Andrea Tirinzoni, Aymen Al Marjani, Emilie Kaufmann
November 27, 2022
November 16, 2022
Kushal Tirumala, Aram H. Markosyan, Armen Aghajanyan, Luke Zettlemoyer
November 16, 2022