You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: _posts/2020-7-20-pytorch-1.6-released.md
+23-56Lines changed: 23 additions & 56 deletions
Original file line number
Diff line number
Diff line change
@@ -4,12 +4,12 @@ title: 'PyTorch 1.6 released w/ Native AMP Support, Microsoft joins as maintaine
4
4
author: Team PyTorch
5
5
---
6
6
7
-
Today, we’re announcing the availability of PyTorch 1.6, along with updated domain libraries. We are also excited to announce the team at <MicrosoftisnowmaintainingWindows> builds and binaries and will also be supporting the community on GitHub as well as the PyTorch Windows [discussion forums](https://discuss.pytorch.org/c/windows/).
7
+
Today, we’re announcing the availability of PyTorch 1.6, along with updated domain libraries. We are also excited to announce the team at <MicrosoftisnowmaintainingWindows> builds and binaries and will also be supporting the community on GitHub as well as the PyTorch Windows discussion forums.
8
8
9
9
The PyTorch 1.6 release includes a number of new APIs, tools for performance improvement and profiling, as well as major updates to both distributed data parallel (DDP) and remote procedure call (RPC) based distributed training.
10
10
A few of the highlights include:
11
11
12
-
1. Automatic mixed precision (AMP) training is now natively supported and a stable feature (See <here> for more details) - thanks for NVIDIA’s contributions;
12
+
1. Automatic mixed precision (AMP) training is now natively supported and a stable feature (See [here] for more details) - thanks for NVIDIA’s contributions;
13
13
2. Native TensorPipe support now added for tensor-aware, point-to-point communication primitives built specifically for machine learning;
14
14
3. Added support for complex tensors to the frontend API surface;
15
15
4. New profiling tools providing tensor-level memory consumption information; and
@@ -23,9 +23,9 @@ Additionally, from this release onward, features will be classified as Stable, B
23
23
24
24
AMP allows users to easily enable automatic mixed precision training enabling higher performance and memory savings of up to 50% on Tensor Core GPUs. Using the natively supported `torch.cuda.amp` API, AMP provides convenience methods for mixed precision, where some operations use the `torch.float32 (float)` datatype and other operations use `torch.float16 (half)`. Some ops, like linear layers and convolutions, are much faster in `float16`. Other ops, like reductions, often require the dynamic range of `float32`. Mixed precision tries to match each op to its appropriate datatype.
@@ -34,7 +34,6 @@ This release adds support for a language-level construct as well as runtime supp
34
34
Parallel execution of TorchScript programs is enabled through two primitives: `torch.jit.fork and torch.jit.wait`. In the below example, we parallelize execution of `foo:`
PyTorch Distributed supports two powerful paradigms: DDP for full sync data parallel training of models and the RPC framework which allows for distributed model parallelism. Currently, these two features work independently and users can’t mix and match these to try out hybrid parallelism paradigms.
107
+
PyTorch Distributed supports two powerful paradigms: DDP for full sync data parallel training of models and the RPC framework which allows for distributed model parallelism. Previously, these two features work independently and users can’t mix and match these to try out hybrid parallelism paradigms.
111
108
112
109
Starting in PyTorch 1.6, we’ve enabled DDP and RPC to work together seamlessly so that users can combine these two techniques to achieve both data parallelism and model parallelism. An example is where users would like to place large embedding tables on parameter servers and use the RPC framework for embedding lookups, but store smaller dense parameters on trainers and use DDP to synchronize the dense parameters. Below is a simple code snippet.
* Tutorial for performant batch RPC using Asynchronous User Functions| Link(https://github.com/pytorch/tutorials/blob/release/1.6/intermediate_source/rpc_async_execution.rst)
* Tutorial for performant batch RPC using Asynchronous User Functions ([Link](https://github.com/pytorch/tutorials/blob/release/1.6/intermediate_source/rpc_async_execution.rst))
@@ -164,15 +160,15 @@ The PyTorch 1.6 release brings beta level support for complex tensors including
164
160
165
161
## torchvision 0.7
166
162
167
-
torchvision 0.7 introduces two new pretrained semantic segmentation models, FCN ResNet50(https://arxiv.org/abs/1411.4038) and DeepLabV3 ResNet50(https://arxiv.org/abs/1706.05587), both trained on COCO and using smaller memory footprints than the ResNet101 backbone. We also introduced support for AMP (Automatic Mixed Precision) autocasting for torchvision models and operators, which automatically selects the floating point precision for different GPU operations to improve performance while maintaining accuracy.
163
+
torchvision 0.7 introduces two new pretrained semantic segmentation models, [FCN ResNet50](https://arxiv.org/abs/1411.4038) and [DeepLabV3 ResNet50](https://arxiv.org/abs/1706.05587), both trained on COCO and using smaller memory footprints than the ResNet101 backbone. We also introduced support for AMP (Automatic Mixed Precision) autocasting for torchvision models and operators, which automatically selects the floating point precision for different GPU operations to improve performance while maintaining accuracy.
torchaudio now officially supports Windows. This release also introduces a new model module (with wav2letter included), new functionals (contrast, cvm, dcshift, overdrive, vad, phaser, flanger, biquad), datasets (GTZAN, CMU), and a new optional sox backend with support for torchscript.
@@ -192,7 +188,7 @@ This is a great opportunity to connect with the community and practice your mach
192
188
193
189
## LPCV Challenge
194
190
195
-
The [2020 CVPR Low-Power Vision Challenge (LPCV) - Online Track for UAV video](https://lpcv.ai/2020CVPR/video-track) submission deadline coming up shortly. You have until July 31, 2020 to build a system that can discover and recognize characters in video captured by an unmanned aerial vehicle (UAV) accurately using PyTorch and Raspberry Pi 3B+.
191
+
The [2020 CVPR Low-Power Vision Challenge (LPCV) - Online Track for UAV video](https://lpcv.ai/2020CVPR/video-track) submission deadline is coming up shortly. You have until July 31, 2020 to build a system that can discover and recognize characters in video captured by an unmanned aerial vehicle (UAV) accurately using PyTorch and Raspberry Pi 3B+.
196
192
197
193
## Prototype Features
198
194
@@ -208,34 +204,5 @@ To reiterate, prototype features in PyTorch are early features that we are looki
0 commit comments