Skip to content

Commit 9873737

Browse files
authored
Update 2020-07-20-accelerating-training-on-nvidia-gpus-with-pytorch-automatic-mixed-precision.md
1 parent 5963817 commit 9873737

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

_posts/2020-07-20-accelerating-training-on-nvidia-gpus-with-pytorch-automatic-mixed-precision.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
---
22
layout: blog_detail
33
title: 'Accelerating training on NVIDIA GPUs with PyTorch automatic mixed precision'
4-
author: Michael Carilli, Mengdi Huang, Chetan Tekur
4+
author: Mengdi Huang, Chetan Tekur, Michael Carilli
55
---
66

77
Most deep learning frameworks, including PyTorch, train with 32-bit floating point (FP32) arithmetic by default. However this is not essential to achieve full accuracy for many deep learning models. In 2017, NVIDIA researchers developed a methodology for [mixed-precision training](https://developer.nvidia.com/blog/mixed-precision-training-deep-neural-networks/), which combined [single-precision](https://blogs.nvidia.com/blog/2019/11/15/whats-the-difference-between-single-double-multi-and-mixed-precision-computing/) (FP32) with half-precision (e.g. FP16) format when training a network, and achieved the same accuracy as FP32 training using the same hyperparameters, with additional performance benefits on NVIDIA GPUs:
@@ -11,7 +11,7 @@ Most deep learning frameworks, including PyTorch, train with 32-bit floating poi
1111

1212
In order to streamline the user experience of training in mixed precision for researchers and practitioners, NVIDIA developed [Apex](https://developer.nvidia.com/blog/apex-pytorch-easy-mixed-precision-training/) in 2018, which is a lightweight PyTorch extension with [Automatic Mixed Precision](https://developer.nvidia.com/automatic-mixed-precision) (AMP) feature. This feature enables automatic conversion of certain GPU operations from FP32 precision to mixed precision, thus improving performance while maintaining accuracy.
1313

14-
For the PyTorch 1.6 release, developers at NVIDIA and Facebook moved mixed precision functionality into PyTorch core as the AMP package, `[torch.cuda.amp](https://pytorch.org/docs/stable/amp.html)`. `torch.cuda.amp` is more flexible and intuitive compared to `apex.amp`. Some of `apex.amp`'s known pain points that `torch.cuda.amp` has been able to fix:
14+
For the PyTorch 1.6 release, developers at NVIDIA and Facebook moved mixed precision functionality into PyTorch core as the AMP package, [torch.cuda.amp](https://pytorch.org/docs/stable/amp.html). `torch.cuda.amp` is more flexible and intuitive compared to `apex.amp`. Some of `apex.amp`'s known pain points that `torch.cuda.amp` has been able to fix:
1515

1616
* Guaranteed PyTorch version compatibility, because it's part of PyTorch
1717
* No need to build extensions

0 commit comments

Comments
 (0)