Skip to content

Commit 93b6017

Browse files
authored
Update and rename 2021-6-15-pytorch-1.9-new-library-releases.md to 2021-6-14-pytorch-1.9-new-library-releases.md
1 parent bec0e5a commit 93b6017

File tree

1 file changed

+20
-12
lines changed

1 file changed

+20
-12
lines changed

_posts/2021-6-15-pytorch-1.9-new-library-releases.md renamed to _posts/2021-6-14-pytorch-1.9-new-library-releases.md

Lines changed: 20 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ Today, we are announcing updates to a number of PyTorch libraries, alongside the
99
Some highlights include:
1010

1111
* **TorchVision** - Added new SSD and SSDLite models, quantized kernels for object detection, GPU Jpeg decoding, and iOS support. See [release notes](https://github.com/pytorch/vision/releases) here.
12-
* **TorchAudio** - Added inference-only wav2vec 2.0 model that can run in non-Python environments (including iOS), improved re-sampling (i.e. Kaiser Window), switched to PyTorch native Complex type, improved filtering and autograd support. See [release notes](https://github.com/pytorch/audio/releases) here.
12+
* **TorchAudio** - Added wav2vec 2.0 model deployable in non-Python environments (including C++, Android, and iOS). Many performance improvements in lfilter, spectral operations, resampling. Added options for quality control in sampling (i.e. Kaiser window support). Initiated the migration of complex tensors operations. Improved autograd support. See [release notes](https://github.com/pytorch/audio/releases) here.
1313
* **TorchText** - Added a new high-performance Vocab module that provides common functional APIs for NLP workflows. See [release notes](https://github.com/pytorch/text/releases) here.
1414

1515
We’d like to thank the community for their support and work on this latest release.
@@ -19,10 +19,18 @@ Features in PyTorch releases are classified as Stable, Beta, and Prototype. You
1919
# TorchVision 0.10
2020

2121
### (Stable) Quantized kernels for object detection
22-
The forward pass of the nms and roi_align operators now support tensors with a quantized dtype, which can help lower the memory footprint of object detection models, particularly on mobile environments. For more details, refer to [the documentation](https://pytorch.org/vision/stable/auto_examples/index.html).
22+
The forward pass of the nms and roi_align operators now support tensors with a quantized dtype, which can help lower the memory footprint of object detection models, particularly on mobile environments. For more details, refer to [the documentation](https://pytorch.org/vision/stable/ops.html#torchvision.ops.roi_align).
23+
24+
### (Stable) Speed optimizations for Tensor transforms
25+
The resize and flip transforms have been optimized and its runtime improved by up to 5x on the CPU.
26+
27+
### (Stable) Documentation improvements
28+
Significant improvements were made to the documentation. In particular, a new gallery of examples is available. These examples visually illustrate how each transform acts on an image, and also properly documents and illustrates the output of the segmentation models.
29+
30+
The example gallery will be extended in the future to provide more comprehensive examples and serve as a reference for common torchvision tasks. For more details, refer to [the documentation](https://pytorch.org/vision/stable/auto_examples/index.html).
2331

2432
### (Beta) New models for detection
25-
SSD and SSDlite are two popular object detection architectures that are efficient in terms of speed and provide good results for low resolution pictures. In this release, we provide implementations for the original SSD model with VGG16 backbone and for its mobile-friendly variant SSDlite with MobileNetV3-Large backbone.
33+
[SSD](https://arxiv.org/abs/1512.02325) and [SSDlite](https://arxiv.org/abs/1801.04381) are two popular object detection architectures that are efficient in terms of speed and provide good results for low resolution pictures. In this release, we provide implementations for the original SSD model with VGG16 backbone and for its mobile-friendly variant SSDlite with MobileNetV3-Large backbone.
2634

2735
The models were pre-trained on COCO train2017 and can be used as follows:
2836

@@ -47,7 +55,7 @@ The following accuracies can be obtained on COCO val2017 (full results available
4755

4856

4957
{:.table.table-striped.table-bordered}
50-
| | Model | mAP | mAP@50 | mAP@75 |
58+
| Model | mAP | mAP@50 | mAP@75 |
5159
| ------------- | ------------- | ------------- | ------------- |
5260
| SSD300 VGG16 | 25.1 | 41.5 | 26.2 |
5361
| SSDlite320 MobileNetV3-Large | 21.3 | 34.3 | 22.1 |
@@ -71,7 +79,7 @@ TorchVision 0.10 now provides pre-compiled iOS binaries for its C++ operators, w
7179
# TorchAudio 0.9.0
7280

7381
### (Stable) Complex Tensor Migration
74-
TorchAudio has functions that handle complex-valued tensors. These functions follow a convention to use an extra dimension to represent real and imaginary parts. (In the following, we call this convention pseudo complex type.) In PyTorch 1.6, the native complex type was introduced. As its API is getting stable, torchaudio has started to migrate to the native complex type.
82+
TorchAudio has functions that handle complex-valued tensors. These functions follow a convention to use an extra dimension to represent real and imaginary parts. In PyTorch 1.6, the native complex type was introduced. As its API is getting stable, torchaudio has started to migrate to the native complex type.
7583

7684
In this release, we added support for native complex tensors, and you can opt-in to use them. Using the native complex types, we have verified that affected functions continue to support autograd and TorchScript, moreover, switching to native complex types improves their performance. For more details, refer to [pytorch/audio#1337](https://github.com/pytorch/audio/issues/1337).
7785

@@ -95,22 +103,22 @@ We have added the model architectures from [Wav2Vec2.0](https://arxiv.org/abs/20
95103
The following code snippet illustrates such a use case. Please check out our [c++ example directory](https://github.com/pytorch/audio/tree/master/examples/libtorchaudio) for the complete example. Currently, it is designed for running inference. If you would like more support for training, please file a feature request.
96104

97105
```python
98-
# Import fine-tuned model from Hugging Face Hub
106+
|# Import fine-tuned model from Hugging Face Hub
99107
import transformers
100108
from torchaudio.models.wav2vec2.utils import import_huggingface_model
101109

102110
original = Wav2Vec2ForCTC.from_pretrained("facebook/wav2vec2-base-960h")
103-
imported = import_huggingface_model(original)
111+
imported = import_huggingface_model(original)|
104112

105-
# Import fine-tuned model from fairseq
113+
|# Import fine-tuned model from fairseq
106114
import fairseq
107115
from torchaudio.models.wav2vec2.utils import import_fairseq_model
108116

109117
original, _, _ = fairseq.checkpoint_utils.load_model_ensemble_and_task(
110118
["wav2vec_small_960h.pt"], arg_overrides={'data': "<data_dir>"})
111-
imported = import_fairseq_model(original[0].w2v_encoder)
119+
imported = import_fairseq_model(original[0].w2v_encoder)|
112120

113-
# Build uninitialized model and load state dict
121+
|# Build uninitialized model and load state dict
114122
from torchaudio.models import wav2vec2_base
115123

116124
model = wav2vec2_base(num_out=32)
@@ -121,7 +129,7 @@ quantized_model = torch.quantization.quantize_dynamic(
121129
model, qconfig_spec={torch.nn.Linear}, dtype=torch.qint8)
122130
scripted_model = torch.jit.script(quantized_model)
123131
optimized_model = optimize_for_mobile(scripted_model)
124-
optimized_model.save("model_for_deployment.pt")
132+
optimized_model.save("model_for_deployment.pt")|
125133
```
126134

127135
For more details, see [the documentation](https://pytorch.org/audio/0.9.0/models.html#wav2vec2-0).
@@ -132,7 +140,7 @@ In release 0.8, we vectorized the operation in ```torchaudio.compliance.kaldi.re
132140
We have:
133141
* Added Kaiser window support for a wider range of resampling quality.
134142
* Added ```rolloff``` parameter for anti-aliasing control.
135-
* Added the mechanism to precompute the kernel and cache it in torchaudio.transforms.Resample for even faster operation.
143+
* Added the mechanism to precompute the kernel and cache it in ```torchaudio.transforms.Resample``` for even faster operation.
136144
* Moved the implementation from ```torchaudio.compliance.kaldi.resample_waveform``` to ```torchaudio.functional.resample``` and deprecated ```torchaudio.compliance.kaldi.resample_waveform```.
137145

138146
For more details, see [the documentation](https://pytorch.org/audio/0.9.0/transforms.html#resample).

0 commit comments

Comments
 (0)