Skip to content

Commit ad66956

Browse files
committed
auto-generating sphinx docs
1 parent 1ebe500 commit ad66956

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

docs/master/_modules/torch/nn/parallel/data_parallel.html

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -373,7 +373,7 @@ <h1>Source code for torch.nn.parallel.data_parallel</h1><div class="highlight"><
373373
<span class="k">return</span>
374374

375375

376-
<div class="viewcode-block" id="DataParallel"><a class="viewcode-back" href="../../../../generated/torch.nn.DataParallel.html#torch.nn.DataParallel">[docs]</a><span class="k">class</span> <span class="nc">DataParallel</span><span class="p">(</span><span class="n">Module</span><span class="p">):</span>
376+
<span class="k">class</span> <span class="nc">DataParallel</span><span class="p">(</span><span class="n">Module</span><span class="p">):</span>
377377
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Implements data parallelism at the module level.</span>
378378

379379
<span class="sd"> This container parallelizes the application of the given :attr:`module` by</span>
@@ -505,10 +505,10 @@ <h1>Source code for torch.nn.parallel.data_parallel</h1><div class="highlight"><
505505
<span class="k">return</span> <span class="n">parallel_apply</span><span class="p">(</span><span class="n">replicas</span><span class="p">,</span> <span class="n">inputs</span><span class="p">,</span> <span class="n">kwargs</span><span class="p">,</span> <span class="bp">self</span><span class="o">.</span><span class="n">device_ids</span><span class="p">[:</span><span class="nb">len</span><span class="p">(</span><span class="n">replicas</span><span class="p">)])</span>
506506

507507
<span class="k">def</span> <span class="nf">gather</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">outputs</span><span class="p">,</span> <span class="n">output_device</span><span class="p">):</span>
508-
<span class="k">return</span> <span class="n">gather</span><span class="p">(</span><span class="n">outputs</span><span class="p">,</span> <span class="n">output_device</span><span class="p">,</span> <span class="n">dim</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dim</span><span class="p">)</span></div>
508+
<span class="k">return</span> <span class="n">gather</span><span class="p">(</span><span class="n">outputs</span><span class="p">,</span> <span class="n">output_device</span><span class="p">,</span> <span class="n">dim</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">dim</span><span class="p">)</span>
509509

510510

511-
<span class="k">def</span> <span class="nf">data_parallel</span><span class="p">(</span><span class="n">module</span><span class="p">,</span> <span class="n">inputs</span><span class="p">,</span> <span class="n">device_ids</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span> <span class="n">output_device</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span> <span class="n">dim</span><span class="o">=</span><span class="mi">0</span><span class="p">,</span> <span class="n">module_kwargs</span><span class="o">=</span><span class="kc">None</span><span class="p">):</span>
511+
<div class="viewcode-block" id="data_parallel"><a class="viewcode-back" href="../../../../nn.functional.html#torch.nn.parallel.data_parallel">[docs]</a><span class="k">def</span> <span class="nf">data_parallel</span><span class="p">(</span><span class="n">module</span><span class="p">,</span> <span class="n">inputs</span><span class="p">,</span> <span class="n">device_ids</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span> <span class="n">output_device</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span> <span class="n">dim</span><span class="o">=</span><span class="mi">0</span><span class="p">,</span> <span class="n">module_kwargs</span><span class="o">=</span><span class="kc">None</span><span class="p">):</span>
512512
<span class="sa">r</span><span class="sd">&quot;&quot;&quot;Evaluates module(input) in parallel across the GPUs given in device_ids.</span>
513513

514514
<span class="sd"> This is the functional version of the DataParallel module.</span>
@@ -548,7 +548,7 @@ <h1>Source code for torch.nn.parallel.data_parallel</h1><div class="highlight"><
548548
<span class="n">used_device_ids</span> <span class="o">=</span> <span class="n">device_ids</span><span class="p">[:</span><span class="nb">len</span><span class="p">(</span><span class="n">inputs</span><span class="p">)]</span>
549549
<span class="n">replicas</span> <span class="o">=</span> <span class="n">replicate</span><span class="p">(</span><span class="n">module</span><span class="p">,</span> <span class="n">used_device_ids</span><span class="p">)</span>
550550
<span class="n">outputs</span> <span class="o">=</span> <span class="n">parallel_apply</span><span class="p">(</span><span class="n">replicas</span><span class="p">,</span> <span class="n">inputs</span><span class="p">,</span> <span class="n">module_kwargs</span><span class="p">,</span> <span class="n">used_device_ids</span><span class="p">)</span>
551-
<span class="k">return</span> <span class="n">gather</span><span class="p">(</span><span class="n">outputs</span><span class="p">,</span> <span class="n">output_device</span><span class="p">,</span> <span class="n">dim</span><span class="p">)</span>
551+
<span class="k">return</span> <span class="n">gather</span><span class="p">(</span><span class="n">outputs</span><span class="p">,</span> <span class="n">output_device</span><span class="p">,</span> <span class="n">dim</span><span class="p">)</span></div>
552552
</pre></div>
553553

554554
</article>

0 commit comments

Comments
 (0)