Skip to content

Commit 92ab92c

Browse files
committed
Generate Python docs from pytorch/pytorch@2540f86
1 parent 7adab3a commit 92ab92c

File tree

1,957 files changed

+2615
-2614
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

1,957 files changed

+2615
-2614
lines changed

docs/master/_images/RReLU.png

257 Bytes
Loading

docs/master/_modules/index.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -217,7 +217,7 @@
217217
<div class="pytorch-left-menu-search">
218218

219219
<div class="version">
220-
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git8c47e9d ) &#x25BC</a>
220+
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git2540f86 ) &#x25BC</a>
221221
</div>
222222

223223

docs/master/_modules/torch.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -217,7 +217,7 @@
217217
<div class="pytorch-left-menu-search">
218218

219219
<div class="version">
220-
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git8c47e9d ) &#x25BC</a>
220+
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git2540f86 ) &#x25BC</a>
221221
</div>
222222

223223

docs/master/_modules/torch/__config__.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -217,7 +217,7 @@
217217
<div class="pytorch-left-menu-search">
218218

219219
<div class="version">
220-
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git8c47e9d ) &#x25BC</a>
220+
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git2540f86 ) &#x25BC</a>
221221
</div>
222222

223223

docs/master/_modules/torch/_jit_internal.html

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -217,7 +217,7 @@
217217
<div class="pytorch-left-menu-search">
218218

219219
<div class="version">
220-
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git8c47e9d ) &#x25BC</a>
220+
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git2540f86 ) &#x25BC</a>
221221
</div>
222222

223223

@@ -932,7 +932,7 @@ <h1>Source code for torch._jit_internal</h1><div class="highlight"><pre>
932932
<span class="k">return</span> <span class="n">fn</span></div>
933933

934934

935-
<span class="k">def</span> <span class="nf">unused</span><span class="p">(</span><span class="n">fn</span><span class="p">):</span>
935+
<div class="viewcode-block" id="unused"><a class="viewcode-back" href="../../generated/torch.jit.unused.html#torch.jit.unused">[docs]</a><span class="k">def</span> <span class="nf">unused</span><span class="p">(</span><span class="n">fn</span><span class="p">):</span>
936936
<span class="sd">&quot;&quot;&quot;</span>
937937
<span class="sd"> This decorator indicates to the compiler that a function or method should</span>
938938
<span class="sd"> be ignored and replaced with the raising of an exception. This allows you</span>
@@ -979,7 +979,7 @@ <h1>Source code for torch._jit_internal</h1><div class="highlight"><pre>
979979
<span class="k">return</span> <span class="n">prop</span>
980980

981981
<span class="n">fn</span><span class="o">.</span><span class="n">_torchscript_modifier</span> <span class="o">=</span> <span class="n">FunctionModifiers</span><span class="o">.</span><span class="n">UNUSED</span>
982-
<span class="k">return</span> <span class="n">fn</span>
982+
<span class="k">return</span> <span class="n">fn</span></div>
983983

984984
<span class="c1"># No op context manager from python side</span>
985985
<span class="k">class</span> <span class="nc">_IgnoreContextManager</span><span class="p">(</span><span class="n">contextlib</span><span class="o">.</span><span class="n">AbstractContextManager</span><span class="p">):</span>
@@ -989,7 +989,7 @@ <h1>Source code for torch._jit_internal</h1><div class="highlight"><pre>
989989
<span class="k">def</span> <span class="fm">__exit__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">exc_type</span><span class="p">:</span> <span class="n">Any</span><span class="p">,</span> <span class="n">exc_value</span><span class="p">:</span> <span class="n">Any</span><span class="p">,</span> <span class="n">traceback</span><span class="p">:</span> <span class="n">Any</span><span class="p">)</span> <span class="o">-&gt;</span> <span class="kc">None</span><span class="p">:</span>
990990
<span class="k">pass</span>
991991

992-
<span class="k">def</span> <span class="nf">ignore</span><span class="p">(</span><span class="n">drop</span><span class="o">=</span><span class="kc">False</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
992+
<div class="viewcode-block" id="ignore"><a class="viewcode-back" href="../../generated/torch.jit.ignore.html#torch.jit.ignore">[docs]</a><span class="k">def</span> <span class="nf">ignore</span><span class="p">(</span><span class="n">drop</span><span class="o">=</span><span class="kc">False</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
993993
<span class="sd">&quot;&quot;&quot;</span>
994994
<span class="sd"> This decorator indicates to the compiler that a function or method should</span>
995995
<span class="sd"> be ignored and left as a Python function. This allows you to leave code in</span>
@@ -1080,7 +1080,7 @@ <h1>Source code for torch._jit_internal</h1><div class="highlight"><pre>
10801080
<span class="k">else</span><span class="p">:</span>
10811081
<span class="n">fn</span><span class="o">.</span><span class="n">_torchscript_modifier</span> <span class="o">=</span> <span class="n">FunctionModifiers</span><span class="o">.</span><span class="n">IGNORE</span>
10821082
<span class="k">return</span> <span class="n">fn</span>
1083-
<span class="k">return</span> <span class="n">decorator</span>
1083+
<span class="k">return</span> <span class="n">decorator</span></div>
10841084

10851085

10861086
<span class="k">def</span> <span class="nf">_copy_to_script_wrapper</span><span class="p">(</span><span class="n">fn</span><span class="p">):</span>

docs/master/_modules/torch/_lobpcg.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -217,7 +217,7 @@
217217
<div class="pytorch-left-menu-search">
218218

219219
<div class="version">
220-
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git8c47e9d ) &#x25BC</a>
220+
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git2540f86 ) &#x25BC</a>
221221
</div>
222222

223223

docs/master/_modules/torch/_lowrank.html

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -217,7 +217,7 @@
217217
<div class="pytorch-left-menu-search">
218218

219219
<div class="version">
220-
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git8c47e9d ) &#x25BC</a>
220+
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git2540f86 ) &#x25BC</a>
221221
</div>
222222

223223

@@ -505,7 +505,7 @@ <h1>Source code for torch._lowrank</h1><div class="highlight"><pre>
505505
<span class="k">return</span> <span class="n">Q</span>
506506

507507

508-
<div class="viewcode-block" id="svd_lowrank"><a class="viewcode-back" href="../../generated/torch.svd_lowrank.html#torch.svd_lowrank">[docs]</a><span class="k">def</span> <span class="nf">svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">:</span> <span class="n">Tensor</span><span class="p">,</span> <span class="n">q</span><span class="p">:</span> <span class="n">Optional</span><span class="p">[</span><span class="nb">int</span><span class="p">]</span> <span class="o">=</span> <span class="mi">6</span><span class="p">,</span> <span class="n">niter</span><span class="p">:</span> <span class="n">Optional</span><span class="p">[</span><span class="nb">int</span><span class="p">]</span> <span class="o">=</span> <span class="mi">2</span><span class="p">,</span>
508+
<span class="k">def</span> <span class="nf">svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">:</span> <span class="n">Tensor</span><span class="p">,</span> <span class="n">q</span><span class="p">:</span> <span class="n">Optional</span><span class="p">[</span><span class="nb">int</span><span class="p">]</span> <span class="o">=</span> <span class="mi">6</span><span class="p">,</span> <span class="n">niter</span><span class="p">:</span> <span class="n">Optional</span><span class="p">[</span><span class="nb">int</span><span class="p">]</span> <span class="o">=</span> <span class="mi">2</span><span class="p">,</span>
509509
<span class="n">M</span><span class="p">:</span> <span class="n">Optional</span><span class="p">[</span><span class="n">Tensor</span><span class="p">]</span> <span class="o">=</span> <span class="kc">None</span><span class="p">)</span> <span class="o">-&gt;</span> <span class="n">Tuple</span><span class="p">[</span><span class="n">Tensor</span><span class="p">,</span> <span class="n">Tensor</span><span class="p">,</span> <span class="n">Tensor</span><span class="p">]:</span>
510510
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Return the singular value decomposition ``(U, S, V)`` of a matrix,</span>
511511
<span class="sd"> batches of matrices, or a sparse matrix :math:`A` such that</span>
@@ -550,7 +550,7 @@ <h1>Source code for torch._lowrank</h1><div class="highlight"><pre>
550550
<span class="n">tensor_ops</span> <span class="o">=</span> <span class="p">(</span><span class="n">A</span><span class="p">,</span> <span class="n">M</span><span class="p">)</span>
551551
<span class="k">if</span> <span class="p">(</span><span class="ow">not</span> <span class="nb">set</span><span class="p">(</span><span class="nb">map</span><span class="p">(</span><span class="nb">type</span><span class="p">,</span> <span class="n">tensor_ops</span><span class="p">))</span><span class="o">.</span><span class="n">issubset</span><span class="p">((</span><span class="n">torch</span><span class="o">.</span><span class="n">Tensor</span><span class="p">,</span> <span class="nb">type</span><span class="p">(</span><span class="kc">None</span><span class="p">)))</span> <span class="ow">and</span> <span class="n">has_torch_function</span><span class="p">(</span><span class="n">tensor_ops</span><span class="p">)):</span>
552552
<span class="k">return</span> <span class="n">handle_torch_function</span><span class="p">(</span><span class="n">svd_lowrank</span><span class="p">,</span> <span class="n">tensor_ops</span><span class="p">,</span> <span class="n">A</span><span class="p">,</span> <span class="n">q</span><span class="o">=</span><span class="n">q</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="n">niter</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="n">M</span><span class="p">)</span>
553-
<span class="k">return</span> <span class="n">_svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">,</span> <span class="n">q</span><span class="o">=</span><span class="n">q</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="n">niter</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="n">M</span><span class="p">)</span></div>
553+
<span class="k">return</span> <span class="n">_svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">,</span> <span class="n">q</span><span class="o">=</span><span class="n">q</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="n">niter</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="n">M</span><span class="p">)</span>
554554

555555

556556
<span class="k">def</span> <span class="nf">_svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">:</span> <span class="n">Tensor</span><span class="p">,</span> <span class="n">q</span><span class="p">:</span> <span class="n">Optional</span><span class="p">[</span><span class="nb">int</span><span class="p">]</span> <span class="o">=</span> <span class="mi">6</span><span class="p">,</span> <span class="n">niter</span><span class="p">:</span> <span class="n">Optional</span><span class="p">[</span><span class="nb">int</span><span class="p">]</span> <span class="o">=</span> <span class="mi">2</span><span class="p">,</span>
@@ -600,7 +600,7 @@ <h1>Source code for torch._lowrank</h1><div class="highlight"><pre>
600600
<span class="k">return</span> <span class="n">U</span><span class="p">,</span> <span class="n">S</span><span class="p">,</span> <span class="n">V</span>
601601

602602

603-
<span class="k">def</span> <span class="nf">pca_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">:</span> <span class="n">Tensor</span><span class="p">,</span> <span class="n">q</span><span class="p">:</span> <span class="n">Optional</span><span class="p">[</span><span class="nb">int</span><span class="p">]</span> <span class="o">=</span> <span class="kc">None</span><span class="p">,</span> <span class="n">center</span><span class="p">:</span> <span class="nb">bool</span> <span class="o">=</span> <span class="kc">True</span><span class="p">,</span>
603+
<div class="viewcode-block" id="pca_lowrank"><a class="viewcode-back" href="../../generated/torch.pca_lowrank.html#torch.pca_lowrank">[docs]</a><span class="k">def</span> <span class="nf">pca_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">:</span> <span class="n">Tensor</span><span class="p">,</span> <span class="n">q</span><span class="p">:</span> <span class="n">Optional</span><span class="p">[</span><span class="nb">int</span><span class="p">]</span> <span class="o">=</span> <span class="kc">None</span><span class="p">,</span> <span class="n">center</span><span class="p">:</span> <span class="nb">bool</span> <span class="o">=</span> <span class="kc">True</span><span class="p">,</span>
604604
<span class="n">niter</span><span class="p">:</span> <span class="nb">int</span> <span class="o">=</span> <span class="mi">2</span><span class="p">)</span> <span class="o">-&gt;</span> <span class="n">Tuple</span><span class="p">[</span><span class="n">Tensor</span><span class="p">,</span> <span class="n">Tensor</span><span class="p">,</span> <span class="n">Tensor</span><span class="p">]:</span>
605605
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Performs linear Principal Component Analysis (PCA) on a low-rank</span>
606606
<span class="sd"> matrix, batches of such matrices, or sparse matrix.</span>
@@ -701,7 +701,7 @@ <h1>Source code for torch._lowrank</h1><div class="highlight"><pre>
701701
<span class="k">return</span> <span class="n">_svd_lowrank</span><span class="p">(</span><span class="n">A</span><span class="p">,</span> <span class="n">q</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="n">niter</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="n">M</span><span class="p">)</span>
702702
<span class="k">else</span><span class="p">:</span>
703703
<span class="n">C</span> <span class="o">=</span> <span class="n">A</span><span class="o">.</span><span class="n">mean</span><span class="p">(</span><span class="n">dim</span><span class="o">=</span><span class="p">(</span><span class="o">-</span><span class="mi">2</span><span class="p">,),</span> <span class="n">keepdim</span><span class="o">=</span><span class="kc">True</span><span class="p">)</span>
704-
<span class="k">return</span> <span class="n">_svd_lowrank</span><span class="p">(</span><span class="n">A</span> <span class="o">-</span> <span class="n">C</span><span class="p">,</span> <span class="n">q</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="n">niter</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="kc">None</span><span class="p">)</span>
704+
<span class="k">return</span> <span class="n">_svd_lowrank</span><span class="p">(</span><span class="n">A</span> <span class="o">-</span> <span class="n">C</span><span class="p">,</span> <span class="n">q</span><span class="p">,</span> <span class="n">niter</span><span class="o">=</span><span class="n">niter</span><span class="p">,</span> <span class="n">M</span><span class="o">=</span><span class="kc">None</span><span class="p">)</span></div>
705705
</pre></div>
706706

707707
</article>

docs/master/_modules/torch/_tensor.html

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -217,7 +217,7 @@
217217
<div class="pytorch-left-menu-search">
218218

219219
<div class="version">
220-
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git8c47e9d ) &#x25BC</a>
220+
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git2540f86 ) &#x25BC</a>
221221
</div>
222222

223223

@@ -950,11 +950,11 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
950950
<span class="k">else</span><span class="p">:</span>
951951
<span class="k">return</span> <span class="bp">self</span><span class="o">.</span><span class="n">flip</span><span class="p">(</span><span class="mi">0</span><span class="p">)</span>
952952

953-
<span class="k">def</span> <span class="nf">norm</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">p</span><span class="o">=</span><span class="s2">&quot;fro&quot;</span><span class="p">,</span> <span class="n">dim</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span> <span class="n">keepdim</span><span class="o">=</span><span class="kc">False</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="kc">None</span><span class="p">):</span>
953+
<div class="viewcode-block" id="Tensor.norm"><a class="viewcode-back" href="../../generated/torch.Tensor.norm.html#torch.Tensor.norm">[docs]</a> <span class="k">def</span> <span class="nf">norm</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">p</span><span class="o">=</span><span class="s2">&quot;fro&quot;</span><span class="p">,</span> <span class="n">dim</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span> <span class="n">keepdim</span><span class="o">=</span><span class="kc">False</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="kc">None</span><span class="p">):</span>
954954
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;See :func:`torch.norm`&quot;&quot;&quot;</span>
955955
<span class="k">if</span> <span class="n">has_torch_function_unary</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span>
956956
<span class="k">return</span> <span class="n">handle_torch_function</span><span class="p">(</span><span class="n">Tensor</span><span class="o">.</span><span class="n">norm</span><span class="p">,</span> <span class="p">(</span><span class="bp">self</span><span class="p">,),</span> <span class="bp">self</span><span class="p">,</span> <span class="n">p</span><span class="o">=</span><span class="n">p</span><span class="p">,</span> <span class="n">dim</span><span class="o">=</span><span class="n">dim</span><span class="p">,</span> <span class="n">keepdim</span><span class="o">=</span><span class="n">keepdim</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">dtype</span><span class="p">)</span>
957-
<span class="k">return</span> <span class="n">torch</span><span class="o">.</span><span class="n">norm</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">p</span><span class="p">,</span> <span class="n">dim</span><span class="p">,</span> <span class="n">keepdim</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">dtype</span><span class="p">)</span>
957+
<span class="k">return</span> <span class="n">torch</span><span class="o">.</span><span class="n">norm</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">p</span><span class="p">,</span> <span class="n">dim</span><span class="p">,</span> <span class="n">keepdim</span><span class="p">,</span> <span class="n">dtype</span><span class="o">=</span><span class="n">dtype</span><span class="p">)</span></div>
958958

959959
<div class="viewcode-block" id="Tensor.lu"><a class="viewcode-back" href="../../generated/torch.Tensor.lu.html#torch.Tensor.lu">[docs]</a> <span class="k">def</span> <span class="nf">lu</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">pivot</span><span class="o">=</span><span class="kc">True</span><span class="p">,</span> <span class="n">get_infos</span><span class="o">=</span><span class="kc">False</span><span class="p">):</span>
960960
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;See :func:`torch.lu`&quot;&quot;&quot;</span>

docs/master/_modules/torch/_tensor_str.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -217,7 +217,7 @@
217217
<div class="pytorch-left-menu-search">
218218

219219
<div class="version">
220-
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git8c47e9d ) &#x25BC</a>
220+
<a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git2540f86 ) &#x25BC</a>
221221
</div>
222222

223223

0 commit comments

Comments
 (0)