Skip to content

Commit ae64c15

Browse files
committed
auto-generating sphinx docs
1 parent 7dde54e commit ae64c15

File tree

721 files changed

+3217
-1146
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

721 files changed

+3217
-1146
lines changed

docs/master/__config__.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -159,7 +159,7 @@
159159

160160

161161
<div class="version">
162-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
162+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
163163
</div>
164164

165165

docs/master/_modules/index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch.html

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

@@ -564,7 +564,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
564564
<span class="k">return</span> <span class="n">module</span> <span class="o">+</span> <span class="n">class_name</span>
565565

566566

567-
<div class="viewcode-block" id="is_tensor"><a class="viewcode-back" href="../generated/torch.is_tensor.html#torch.is_tensor">[docs]</a><span class="k">def</span> <span class="nf">is_tensor</span><span class="p">(</span><span class="n">obj</span><span class="p">):</span>
567+
<span class="k">def</span> <span class="nf">is_tensor</span><span class="p">(</span><span class="n">obj</span><span class="p">):</span>
568568
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Returns True if `obj` is a PyTorch tensor.</span>
569569

570570
<span class="sd"> Note that this function is simply doing ``isinstance(obj, Tensor)``.</span>
@@ -575,16 +575,16 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
575575
<span class="sd"> Args:</span>
576576
<span class="sd"> obj (Object): Object to test</span>
577577
<span class="sd"> &quot;&quot;&quot;</span>
578-
<span class="k">return</span> <span class="nb">isinstance</span><span class="p">(</span><span class="n">obj</span><span class="p">,</span> <span class="n">torch</span><span class="o">.</span><span class="n">Tensor</span><span class="p">)</span></div>
578+
<span class="k">return</span> <span class="nb">isinstance</span><span class="p">(</span><span class="n">obj</span><span class="p">,</span> <span class="n">torch</span><span class="o">.</span><span class="n">Tensor</span><span class="p">)</span>
579579

580580

581-
<div class="viewcode-block" id="is_storage"><a class="viewcode-back" href="../generated/torch.is_storage.html#torch.is_storage">[docs]</a><span class="k">def</span> <span class="nf">is_storage</span><span class="p">(</span><span class="n">obj</span><span class="p">):</span>
581+
<span class="k">def</span> <span class="nf">is_storage</span><span class="p">(</span><span class="n">obj</span><span class="p">):</span>
582582
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Returns True if `obj` is a PyTorch storage object.</span>
583583

584584
<span class="sd"> Args:</span>
585585
<span class="sd"> obj (Object): Object to test</span>
586586
<span class="sd"> &quot;&quot;&quot;</span>
587-
<span class="k">return</span> <span class="nb">type</span><span class="p">(</span><span class="n">obj</span><span class="p">)</span> <span class="ow">in</span> <span class="n">_storage_classes</span></div>
587+
<span class="k">return</span> <span class="nb">type</span><span class="p">(</span><span class="n">obj</span><span class="p">)</span> <span class="ow">in</span> <span class="n">_storage_classes</span>
588588

589589

590590
<span class="k">def</span> <span class="nf">set_default_tensor_type</span><span class="p">(</span><span class="n">t</span><span class="p">):</span>
@@ -674,8 +674,8 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
674674
<span class="k">pass</span>
675675

676676

677-
<span class="k">class</span> <span class="nc">FloatStorage</span><span class="p">(</span><span class="n">_C</span><span class="o">.</span><span class="n">FloatStorageBase</span><span class="p">,</span> <span class="n">_StorageBase</span><span class="p">):</span>
678-
<span class="k">pass</span>
677+
<div class="viewcode-block" id="FloatStorage"><a class="viewcode-back" href="../storage.html#torch.FloatStorage">[docs]</a><span class="k">class</span> <span class="nc">FloatStorage</span><span class="p">(</span><span class="n">_C</span><span class="o">.</span><span class="n">FloatStorageBase</span><span class="p">,</span> <span class="n">_StorageBase</span><span class="p">):</span>
678+
<span class="k">pass</span></div>
679679

680680

681681
<span class="k">class</span> <span class="nc">HalfStorage</span><span class="p">(</span><span class="n">_C</span><span class="o">.</span><span class="n">HalfStorageBase</span><span class="p">,</span> <span class="n">_StorageBase</span><span class="p">):</span>

docs/master/_modules/torch/__config__.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/_jit_internal.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/_lobpcg.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/_lowrank.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/_tensor_str.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/_utils.html

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

@@ -528,11 +528,15 @@ <h1>Source code for torch._utils</h1><div class="highlight"><pre>
528528
<span class="k">if</span> <span class="n">qscheme</span> <span class="o">==</span> <span class="n">torch</span><span class="o">.</span><span class="n">per_tensor_affine</span><span class="p">:</span>
529529
<span class="n">_</span><span class="p">,</span> <span class="n">scale</span><span class="p">,</span> <span class="n">zero_point</span> <span class="o">=</span> <span class="n">quantizer_params</span>
530530
<span class="n">tensor</span> <span class="o">=</span> <span class="n">torch</span><span class="o">.</span><span class="n">_empty_affine_quantized</span><span class="p">(</span><span class="n">size</span><span class="p">,</span> <span class="n">scale</span><span class="o">=</span><span class="n">scale</span><span class="p">,</span> <span class="n">zero_point</span><span class="o">=</span><span class="n">zero_point</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">storage</span><span class="o">.</span><span class="n">dtype</span><span class="p">)</span>
531-
<span class="k">elif</span> <span class="n">qscheme</span> <span class="o">==</span> <span class="n">torch</span><span class="o">.</span><span class="n">per_channel_affine</span><span class="p">:</span>
531+
<span class="k">elif</span> <span class="n">qscheme</span> <span class="ow">in</span> <span class="p">(</span><span class="n">torch</span><span class="o">.</span><span class="n">per_channel_affine</span><span class="p">,</span> <span class="n">torch</span><span class="o">.</span><span class="n">per_channel_affine_float_qparams</span><span class="p">):</span>
532532
<span class="n">_</span><span class="p">,</span> <span class="n">scales</span><span class="p">,</span> <span class="n">zero_points</span><span class="p">,</span> <span class="n">axis</span> <span class="o">=</span> <span class="n">quantizer_params</span>
533533
<span class="k">if</span> <span class="nb">type</span><span class="p">(</span><span class="n">scales</span><span class="p">)</span> <span class="ow">is</span> <span class="nb">list</span> <span class="ow">and</span> <span class="nb">type</span><span class="p">(</span><span class="n">zero_points</span><span class="p">)</span> <span class="ow">is</span> <span class="nb">list</span><span class="p">:</span>
534-
<span class="n">scales</span> <span class="o">=</span> <span class="n">torch</span><span class="o">.</span><span class="n">tensor</span><span class="p">(</span><span class="n">scales</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">torch</span><span class="o">.</span><span class="n">double</span><span class="p">)</span>
535-
<span class="n">zero_points</span> <span class="o">=</span> <span class="n">torch</span><span class="o">.</span><span class="n">tensor</span><span class="p">(</span><span class="n">zero_points</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">torch</span><span class="o">.</span><span class="n">long</span><span class="p">)</span>
534+
<span class="k">if</span> <span class="n">qscheme</span> <span class="o">==</span> <span class="n">torch</span><span class="o">.</span><span class="n">per_channel_affine</span><span class="p">:</span>
535+
<span class="n">scales</span> <span class="o">=</span> <span class="n">torch</span><span class="o">.</span><span class="n">tensor</span><span class="p">(</span><span class="n">scales</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">torch</span><span class="o">.</span><span class="n">double</span><span class="p">)</span>
536+
<span class="n">zero_points</span> <span class="o">=</span> <span class="n">torch</span><span class="o">.</span><span class="n">tensor</span><span class="p">(</span><span class="n">zero_points</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">torch</span><span class="o">.</span><span class="n">long</span><span class="p">)</span>
537+
<span class="k">else</span><span class="p">:</span>
538+
<span class="n">scales</span> <span class="o">=</span> <span class="n">torch</span><span class="o">.</span><span class="n">tensor</span><span class="p">(</span><span class="n">scales</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">torch</span><span class="o">.</span><span class="n">float</span><span class="p">)</span>
539+
<span class="n">zero_points</span> <span class="o">=</span> <span class="n">torch</span><span class="o">.</span><span class="n">tensor</span><span class="p">(</span><span class="n">zero_points</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">torch</span><span class="o">.</span><span class="n">float</span><span class="p">)</span>
536540
<span class="n">tensor</span> <span class="o">=</span> <span class="n">torch</span><span class="o">.</span><span class="n">_empty_per_channel_affine_quantized</span><span class="p">(</span>
537541
<span class="n">size</span><span class="p">,</span> <span class="n">scales</span><span class="o">=</span><span class="n">scales</span><span class="p">,</span> <span class="n">zero_points</span><span class="o">=</span><span class="n">zero_points</span><span class="p">,</span> <span class="n">axis</span><span class="o">=</span><span class="n">axis</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">storage</span><span class="o">.</span><span class="n">dtype</span><span class="p">)</span>
538542
<span class="k">else</span><span class="p">:</span>

docs/master/_modules/torch/autograd.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/autograd/anomaly_mode.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/autograd/function.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/autograd/functional.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/autograd/grad_mode.html

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

@@ -376,7 +376,7 @@ <h1>Source code for torch.autograd.grad_mode</h1><div class="highlight"><pre>
376376
<span class="k">return</span> <span class="n">generator_context</span>
377377

378378

379-
<div class="viewcode-block" id="no_grad"><a class="viewcode-back" href="../../../autograd.html#torch.autograd.no_grad">[docs]</a><span class="k">class</span> <span class="nc">no_grad</span><span class="p">(</span><span class="n">_DecoratorContextManager</span><span class="p">):</span>
379+
<span class="k">class</span> <span class="nc">no_grad</span><span class="p">(</span><span class="n">_DecoratorContextManager</span><span class="p">):</span>
380380
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Context-manager that disabled gradient calculation.</span>
381381

382382
<span class="sd"> Disabling gradient calculation is useful for inference, when you are sure</span>
@@ -411,10 +411,10 @@ <h1>Source code for torch.autograd.grad_mode</h1><div class="highlight"><pre>
411411
<span class="n">torch</span><span class="o">.</span><span class="n">_C</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="kc">False</span><span class="p">)</span>
412412

413413
<span class="k">def</span> <span class="fm">__exit__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="o">*</span><span class="n">args</span><span class="p">):</span>
414-
<span class="n">torch</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">prev</span><span class="p">)</span></div>
414+
<span class="n">torch</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">prev</span><span class="p">)</span>
415415

416416

417-
<div class="viewcode-block" id="enable_grad"><a class="viewcode-back" href="../../../autograd.html#torch.autograd.enable_grad">[docs]</a><span class="k">class</span> <span class="nc">enable_grad</span><span class="p">(</span><span class="n">_DecoratorContextManager</span><span class="p">):</span>
417+
<div class="viewcode-block" id="enable_grad"><a class="viewcode-back" href="../../../generated/torch.enable_grad.html#torch.enable_grad">[docs]</a><span class="k">class</span> <span class="nc">enable_grad</span><span class="p">(</span><span class="n">_DecoratorContextManager</span><span class="p">):</span>
418418
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Context-manager that enables gradient calculation.</span>
419419

420420
<span class="sd"> Enables gradient calculation, if it has been disabled via :class:`~no_grad`</span>
@@ -453,7 +453,7 @@ <h1>Source code for torch.autograd.grad_mode</h1><div class="highlight"><pre>
453453
<span class="n">torch</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">prev</span><span class="p">)</span></div>
454454

455455

456-
<div class="viewcode-block" id="set_grad_enabled"><a class="viewcode-back" href="../../../autograd.html#torch.autograd.set_grad_enabled">[docs]</a><span class="k">class</span> <span class="nc">set_grad_enabled</span><span class="p">(</span><span class="nb">object</span><span class="p">):</span>
456+
<span class="k">class</span> <span class="nc">set_grad_enabled</span><span class="p">(</span><span class="nb">object</span><span class="p">):</span>
457457
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Context-manager that sets gradient calculation to on or off.</span>
458458

459459
<span class="sd"> ``set_grad_enabled`` will enable or disable grads based on its argument :attr:`mode`.</span>
@@ -495,7 +495,7 @@ <h1>Source code for torch.autograd.grad_mode</h1><div class="highlight"><pre>
495495
<span class="k">pass</span>
496496

497497
<span class="k">def</span> <span class="fm">__exit__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="o">*</span><span class="n">args</span><span class="p">):</span>
498-
<span class="n">torch</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">prev</span><span class="p">)</span></div>
498+
<span class="n">torch</span><span class="o">.</span><span class="n">set_grad_enabled</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">prev</span><span class="p">)</span>
499499
</pre></div>
500500

501501
</article>

docs/master/_modules/torch/autograd/gradcheck.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/autograd/profiler.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/cuda.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/cuda/amp/autocast_mode.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+d73c26d &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+bdf9700 &#x25BC</a>
162162
</div>
163163

164164

0 commit comments

Comments
 (0)