Skip to content

Commit db37b8c

Browse files
committed
auto-generating sphinx docs
1 parent dea55bc commit db37b8c

File tree

748 files changed

+2565
-898
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

748 files changed

+2565
-898
lines changed

docs/master/__config__.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -159,7 +159,7 @@
159159

160160

161161
<div class="version">
162-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+69d74c8 &#x25BC</a>
162+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+970389e &#x25BC</a>
163163
</div>
164164

165165

docs/master/_modules/index.html

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+69d74c8 &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+970389e &#x25BC</a>
162162
</div>
163163

164164

@@ -350,6 +350,7 @@ <h1>All modules for which code is available</h1>
350350
<li><a href="torch/_lowrank.html">torch._lowrank</a></li>
351351
<li><a href="torch/_tensor_str.html">torch._tensor_str</a></li>
352352
<li><a href="torch/_utils.html">torch._utils</a></li>
353+
<li><a href="torch/_vmap_internals.html">torch._vmap_internals</a></li>
353354
<li><a href="torch/autograd.html">torch.autograd</a></li>
354355
<ul><li><a href="torch/autograd/anomaly_mode.html">torch.autograd.anomaly_mode</a></li>
355356
<li><a href="torch/autograd/function.html">torch.autograd.function</a></li>

docs/master/_modules/torch.html

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+69d74c8 &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+970389e &#x25BC</a>
162162
</div>
163163

164164

@@ -588,7 +588,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
588588
<span class="k">return</span> <span class="nb">type</span><span class="p">(</span><span class="n">obj</span><span class="p">)</span> <span class="ow">in</span> <span class="n">_storage_classes</span></div>
589589

590590

591-
<div class="viewcode-block" id="set_default_tensor_type"><a class="viewcode-back" href="../generated/torch.set_default_tensor_type.html#torch.set_default_tensor_type">[docs]</a><span class="k">def</span> <span class="nf">set_default_tensor_type</span><span class="p">(</span><span class="n">t</span><span class="p">):</span>
591+
<span class="k">def</span> <span class="nf">set_default_tensor_type</span><span class="p">(</span><span class="n">t</span><span class="p">):</span>
592592
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Sets the default ``torch.Tensor`` type to floating point tensor type</span>
593593
<span class="sd"> ``t``. This type will also be used as default floating point type for</span>
594594
<span class="sd"> type inference in :func:`torch.tensor`.</span>
@@ -609,10 +609,10 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
609609
<span class="sd"> &quot;&quot;&quot;</span>
610610
<span class="k">if</span> <span class="nb">isinstance</span><span class="p">(</span><span class="n">t</span><span class="p">,</span> <span class="n">_string_classes</span><span class="p">):</span>
611611
<span class="n">t</span> <span class="o">=</span> <span class="n">_import_dotted_name</span><span class="p">(</span><span class="n">t</span><span class="p">)</span>
612-
<span class="n">_C</span><span class="o">.</span><span class="n">_set_default_tensor_type</span><span class="p">(</span><span class="n">t</span><span class="p">)</span></div>
612+
<span class="n">_C</span><span class="o">.</span><span class="n">_set_default_tensor_type</span><span class="p">(</span><span class="n">t</span><span class="p">)</span>
613613

614614

615-
<div class="viewcode-block" id="set_default_dtype"><a class="viewcode-back" href="../generated/torch.set_default_dtype.html#torch.set_default_dtype">[docs]</a><span class="k">def</span> <span class="nf">set_default_dtype</span><span class="p">(</span><span class="n">d</span><span class="p">):</span>
615+
<span class="k">def</span> <span class="nf">set_default_dtype</span><span class="p">(</span><span class="n">d</span><span class="p">):</span>
616616
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Sets the default floating point dtype to :attr:`d`.</span>
617617
<span class="sd"> This dtype is:</span>
618618

@@ -640,9 +640,9 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
640640
<span class="sd"> torch.complex128</span>
641641

642642
<span class="sd"> &quot;&quot;&quot;</span>
643-
<span class="n">_C</span><span class="o">.</span><span class="n">_set_default_dtype</span><span class="p">(</span><span class="n">d</span><span class="p">)</span></div>
643+
<span class="n">_C</span><span class="o">.</span><span class="n">_set_default_dtype</span><span class="p">(</span><span class="n">d</span><span class="p">)</span>
644644

645-
<div class="viewcode-block" id="set_deterministic"><a class="viewcode-back" href="../generated/torch.set_deterministic.html#torch.set_deterministic">[docs]</a><span class="k">def</span> <span class="nf">set_deterministic</span><span class="p">(</span><span class="n">d</span><span class="p">):</span>
645+
<span class="k">def</span> <span class="nf">set_deterministic</span><span class="p">(</span><span class="n">d</span><span class="p">):</span>
646646
<span class="sa">r</span><span class="sd">&quot;&quot;&quot; Sets whether native PyTorch operations must use deterministic</span>
647647
<span class="sd"> algorithms. When True, operations without deterministic algorithms</span>
648648
<span class="sd"> will throw a :class:RuntimeError when called.</span>
@@ -712,7 +712,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
712712
<span class="sd"> d (:class:`bool`): If True, force operations to be deterministic.</span>
713713
<span class="sd"> If False, allow non-deterministic operations.</span>
714714
<span class="sd"> &quot;&quot;&quot;</span>
715-
<span class="n">_C</span><span class="o">.</span><span class="n">_set_deterministic</span><span class="p">(</span><span class="n">d</span><span class="p">)</span></div>
715+
<span class="n">_C</span><span class="o">.</span><span class="n">_set_deterministic</span><span class="p">(</span><span class="n">d</span><span class="p">)</span>
716716

717717
<div class="viewcode-block" id="is_deterministic"><a class="viewcode-back" href="../generated/torch.is_deterministic.html#torch.is_deterministic">[docs]</a><span class="k">def</span> <span class="nf">is_deterministic</span><span class="p">():</span>
718718
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Returns True if the global deterministic flag is turned on. Refer to</span>
@@ -896,9 +896,9 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
896896
<span class="k">del</span> <span class="n">_torch_docs</span><span class="p">,</span> <span class="n">_tensor_docs</span><span class="p">,</span> <span class="n">_storage_docs</span>
897897

898898

899-
<div class="viewcode-block" id="compiled_with_cxx11_abi"><a class="viewcode-back" href="../generated/torch.compiled_with_cxx11_abi.html#torch.compiled_with_cxx11_abi">[docs]</a><span class="k">def</span> <span class="nf">compiled_with_cxx11_abi</span><span class="p">():</span>
899+
<span class="k">def</span> <span class="nf">compiled_with_cxx11_abi</span><span class="p">():</span>
900900
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Returns whether PyTorch was built with _GLIBCXX_USE_CXX11_ABI=1&quot;&quot;&quot;</span>
901-
<span class="k">return</span> <span class="n">_C</span><span class="o">.</span><span class="n">_GLIBCXX_USE_CXX11_ABI</span></div>
901+
<span class="k">return</span> <span class="n">_C</span><span class="o">.</span><span class="n">_GLIBCXX_USE_CXX11_ABI</span>
902902

903903

904904
<span class="c1"># Import the ops &quot;namespace&quot;</span>

docs/master/_modules/torch/__config__.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+69d74c8 &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+970389e &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/_jit_internal.html

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+69d74c8 &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+970389e &#x25BC</a>
162162
</div>
163163

164164

@@ -1083,7 +1083,7 @@ <h1>Source code for torch._jit_internal</h1><div class="highlight"><pre>
10831083
<span class="nb">globals</span><span class="p">()[</span><span class="sa">f</span><span class="s2">&quot;BroadcastingList</span><span class="si">{</span><span class="n">i</span><span class="si">}</span><span class="s2">&quot;</span><span class="p">]</span> <span class="o">=</span> <span class="n">BroadcastingList1</span>
10841084

10851085

1086-
<div class="viewcode-block" id="is_scripting"><a class="viewcode-back" href="../../jit_language_reference.html#torch.jit.is_scripting">[docs]</a><span class="k">def</span> <span class="nf">is_scripting</span><span class="p">():</span>
1086+
<span class="k">def</span> <span class="nf">is_scripting</span><span class="p">():</span>
10871087
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;</span>
10881088
<span class="sd"> Function that returns True when in compilation and False otherwise. This</span>
10891089
<span class="sd"> is useful especially with the @unused decorator to leave code in your</span>
@@ -1102,7 +1102,7 @@ <h1>Source code for torch._jit_internal</h1><div class="highlight"><pre>
11021102
<span class="sd"> else:</span>
11031103
<span class="sd"> return unsupported_linear_op(x)</span>
11041104
<span class="sd"> &quot;&quot;&quot;</span>
1105-
<span class="k">return</span> <span class="kc">False</span></div>
1105+
<span class="k">return</span> <span class="kc">False</span>
11061106

11071107

11081108
<span class="c1"># Retrieves a fully-qualified name (module hierarchy + classname) for a given obj.</span>

docs/master/_modules/torch/_lobpcg.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+69d74c8 &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+970389e &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/_lowrank.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+69d74c8 &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+970389e &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/_tensor_str.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+69d74c8 &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+970389e &#x25BC</a>
162162
</div>
163163

164164

docs/master/_modules/torch/_utils.html

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -158,7 +158,7 @@
158158

159159

160160
<div class="version">
161-
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+69d74c8 &#x25BC</a>
161+
<a href='http://pytorch.org/docs/versions.html'>1.7.0a0+970389e &#x25BC</a>
162162
</div>
163163

164164

0 commit comments

Comments
 (0)