|
207 | 207 | <div class="pytorch-left-menu-search"> |
208 | 208 |
|
209 | 209 | <div class="version"> |
210 | | - <a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git52c516e ) ▼</a> |
| 210 | + <a href='https://pytorch.org/docs/versions.html'>master (1.12.0a0+git08889b2 ) ▼</a> |
211 | 211 | </div> |
212 | 212 |
|
213 | 213 |
|
@@ -519,9 +519,9 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre> |
519 | 519 | <span class="k">else</span><span class="p">:</span> |
520 | 520 | <span class="k">raise</span> <span class="ne">RuntimeError</span><span class="p">(</span><span class="sa">f</span><span class="s2">"Unsupported qscheme </span><span class="si">{</span><span class="bp">self</span><span class="o">.</span><span class="n">qscheme</span><span class="p">()</span><span class="si">}</span><span class="s2"> in deepcopy"</span><span class="p">)</span> |
521 | 521 | <span class="c1"># TODO: Once we decide to break serialization FC, no longer</span> |
522 | | - <span class="c1"># need to wrap with TypedStorage</span> |
| 522 | + <span class="c1"># need to wrap with _TypedStorage</span> |
523 | 523 | <span class="n">new_tensor</span> <span class="o">=</span> <span class="n">torch</span><span class="o">.</span><span class="n">_utils</span><span class="o">.</span><span class="n">_rebuild_qtensor</span><span class="p">(</span> |
524 | | - <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">TypedStorage</span><span class="p">(</span> |
| 524 | + <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">_TypedStorage</span><span class="p">(</span> |
525 | 525 | <span class="n">wrap_storage</span><span class="o">=</span><span class="n">new_storage</span><span class="o">.</span><span class="n">_untyped</span><span class="p">(),</span> |
526 | 526 | <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span><span class="p">),</span> |
527 | 527 | <span class="bp">self</span><span class="o">.</span><span class="n">storage_offset</span><span class="p">(),</span> |
@@ -642,9 +642,9 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre> |
642 | 642 | <span class="k">else</span><span class="p">:</span> |
643 | 643 | <span class="k">raise</span> <span class="ne">RuntimeError</span><span class="p">(</span><span class="sa">f</span><span class="s2">"Serialization is not supported for tensors of type </span><span class="si">{</span><span class="bp">self</span><span class="o">.</span><span class="n">qscheme</span><span class="p">()</span><span class="si">}</span><span class="s2">"</span><span class="p">)</span> |
644 | 644 | <span class="c1"># TODO: Once we decide to break serialization FC, no longer</span> |
645 | | - <span class="c1"># need to wrap with TypedStorage</span> |
| 645 | + <span class="c1"># need to wrap with _TypedStorage</span> |
646 | 646 | <span class="n">args_qtensor</span> <span class="o">=</span> <span class="p">(</span> |
647 | | - <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">TypedStorage</span><span class="p">(</span> |
| 647 | + <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">_TypedStorage</span><span class="p">(</span> |
648 | 648 | <span class="n">wrap_storage</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">storage</span><span class="p">()</span><span class="o">.</span><span class="n">_untyped</span><span class="p">(),</span> |
649 | 649 | <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span><span class="p">),</span> |
650 | 650 | <span class="bp">self</span><span class="o">.</span><span class="n">storage_offset</span><span class="p">(),</span> |
@@ -677,9 +677,9 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre> |
677 | 677 | <span class="k">return</span> <span class="p">(</span><span class="n">torch</span><span class="o">.</span><span class="n">_utils</span><span class="o">.</span><span class="n">_rebuild_sparse_csr_tensor</span><span class="p">,</span> <span class="n">args_sparse_csr</span><span class="p">)</span> |
678 | 678 | <span class="k">else</span><span class="p">:</span> |
679 | 679 | <span class="c1"># TODO: Once we decide to break serialization FC, no longer</span> |
680 | | - <span class="c1"># need to wrap with TypedStorage</span> |
| 680 | + <span class="c1"># need to wrap with _TypedStorage</span> |
681 | 681 | <span class="n">args</span> <span class="o">=</span> <span class="p">(</span> |
682 | | - <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">TypedStorage</span><span class="p">(</span> |
| 682 | + <span class="n">torch</span><span class="o">.</span><span class="n">storage</span><span class="o">.</span><span class="n">_TypedStorage</span><span class="p">(</span> |
683 | 683 | <span class="n">wrap_storage</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">storage</span><span class="p">()</span><span class="o">.</span><span class="n">_untyped</span><span class="p">(),</span> |
684 | 684 | <span class="n">dtype</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dtype</span><span class="p">),</span> |
685 | 685 | <span class="bp">self</span><span class="o">.</span><span class="n">storage_offset</span><span class="p">(),</span> |
@@ -1240,9 +1240,9 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre> |
1240 | 1240 | <span class="sd"> Returns the type of the underlying storage.</span> |
1241 | 1241 |
|
1242 | 1242 | <span class="sd"> """</span> |
1243 | | - <span class="c1"># NB: this returns old fashioned TypedStorage, e.g., FloatStorage, as it</span> |
| 1243 | + <span class="c1"># NB: this returns old fashioned _TypedStorage, e.g., FloatStorage, as it</span> |
1244 | 1244 | <span class="c1"># would be pretty pointless otherwise (it would always return</span> |
1245 | | - <span class="c1"># UntypedStorage)</span> |
| 1245 | + <span class="c1"># _UntypedStorage)</span> |
1246 | 1246 | <span class="k">return</span> <span class="nb">type</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">storage</span><span class="p">())</span> |
1247 | 1247 |
|
1248 | 1248 | <span class="k">def</span> <span class="nf">refine_names</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="o">*</span><span class="n">names</span><span class="p">):</span> |
|
0 commit comments