@@ -849,7 +849,7 @@ <h2 id="B">B</h2>
849
849
< li > < a href ="autograd.html#torch.autograd.backward "> backward() (in module torch.autograd)</ a >
850
850
851
851
< ul >
852
- < li > < a href ="rpc/rpc .html#torch.distributed.autograd.backward "> (in module torch.distributed.autograd)</ a >
852
+ < li > < a href ="rpc.html#torch.distributed.autograd.backward "> (in module torch.distributed.autograd)</ a >
853
853
</ li >
854
854
< li > < a href ="autograd.html#torch.autograd.Function.backward "> (torch.autograd.Function static method)</ a >
855
855
</ li >
@@ -1188,7 +1188,7 @@ <h2 id="C">C</h2>
1188
1188
</ li >
1189
1189
< li > < a href ="distributions.html#torch.distributions.beta.Beta.concentration1 "> concentration1() (torch.distributions.beta.Beta property)</ a >
1190
1190
</ li >
1191
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.RRef.confirmed_by_owner "> confirmed_by_owner() (torch.distributed.rpc.RRef method)</ a >
1191
+ < li > < a href ="rpc.html#torch.distributed.rpc.RRef.confirmed_by_owner "> confirmed_by_owner() (torch.distributed.rpc.RRef method)</ a >
1192
1192
</ li >
1193
1193
< li > < a href ="torch.html#torch.conj "> conj() (in module torch)</ a >
1194
1194
@@ -1210,7 +1210,7 @@ <h2 id="C">C</h2>
1210
1210
</ li >
1211
1211
< li > < a href ="distributions.html#torch.distributions.constraint_registry.ConstraintRegistry "> ConstraintRegistry (class in torch.distributions.constraint_registry)</ a >
1212
1212
</ li >
1213
- < li > < a href ="rpc/rpc .html#torch.distributed.autograd.context "> context (class in torch.distributed.autograd)</ a >
1213
+ < li > < a href ="rpc.html#torch.distributed.autograd.context "> context (class in torch.distributed.autograd)</ a >
1214
1214
</ li >
1215
1215
< li > < a href ="tensors.html#torch.Tensor.contiguous "> contiguous() (torch.Tensor method)</ a >
1216
1216
</ li >
@@ -1536,7 +1536,7 @@ <h2 id="D">D</h2>
1536
1536
</ ul > </ li >
1537
1537
< li > < a href ="nn.html#torch.nn.parallel.DistributedDataParallel "> DistributedDataParallel (class in torch.nn.parallel)</ a >
1538
1538
</ li >
1539
- < li > < a href ="rpc/rpc .html#torch.distributed.optim.DistributedOptimizer "> DistributedOptimizer (class in torch.distributed.optim)</ a >
1539
+ < li > < a href ="rpc.html#torch.distributed.optim.DistributedOptimizer "> DistributedOptimizer (class in torch.distributed.optim)</ a >
1540
1540
</ li >
1541
1541
< li > < a href ="data.html#torch.utils.data.distributed.DistributedSampler "> DistributedSampler (class in torch.utils.data.distributed)</ a >
1542
1542
</ li >
@@ -2175,7 +2175,7 @@ <h2 id="G">G</h2>
2175
2175
</ li >
2176
2176
< li > < a href ="cuda.html#torch.cuda.get_device_name "> get_device_name() (in module torch.cuda)</ a >
2177
2177
</ li >
2178
- < li > < a href ="rpc/rpc .html#torch.distributed.autograd.get_gradients "> get_gradients() (in module torch.distributed.autograd)</ a >
2178
+ < li > < a href ="rpc.html#torch.distributed.autograd.get_gradients "> get_gradients() (in module torch.distributed.autograd)</ a >
2179
2179
</ li >
2180
2180
< li > < a href ="amp.html#torch.cuda.amp.GradScaler.get_growth_factor "> get_growth_factor() (torch.cuda.amp.GradScaler method)</ a >
2181
2181
</ li >
@@ -2211,7 +2211,7 @@ <h2 id="G">G</h2>
2211
2211
</ li >
2212
2212
< li > < a href ="torch.html#torch._C.Generator.get_state "> get_state() (torch._C.Generator method)</ a >
2213
2213
</ li >
2214
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.get_worker_info "> get_worker_info() (in module torch.distributed.rpc)</ a >
2214
+ < li > < a href ="rpc.html#torch.distributed.rpc.get_worker_info "> get_worker_info() (in module torch.distributed.rpc)</ a >
2215
2215
2216
2216
< ul >
2217
2217
< li > < a href ="data.html#torch.utils.data.get_worker_info "> (in module torch.utils.data)</ a >
@@ -2429,7 +2429,7 @@ <h2 id="I">I</h2>
2429
2429
< li > < a href ="distributions.html#torch.distributions.uniform.Uniform.icdf "> (torch.distributions.uniform.Uniform method)</ a >
2430
2430
</ li >
2431
2431
</ ul > </ li >
2432
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.WorkerInfo.id "> id() (torch.distributed.rpc.WorkerInfo property)</ a >
2432
+ < li > < a href ="rpc.html#torch.distributed.rpc.WorkerInfo.id "> id() (torch.distributed.rpc.WorkerInfo property)</ a >
2433
2433
</ li >
2434
2434
< li > < a href ="nn.html#torch.nn.Identity "> Identity (class in torch.nn)</ a >
2435
2435
@@ -2489,11 +2489,11 @@ <h2 id="I">I</h2>
2489
2489
</ li >
2490
2490
< li > < a href ="cuda.html#torch.cuda.init "> init() (in module torch.cuda)</ a >
2491
2491
</ li >
2492
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.ProcessGroupRpcBackendOptions.init_method "> init_method() (torch.distributed.rpc.ProcessGroupRpcBackendOptions property)</ a >
2492
+ < li > < a href ="rpc.html#torch.distributed.rpc.ProcessGroupRpcBackendOptions.init_method "> init_method() (torch.distributed.rpc.ProcessGroupRpcBackendOptions property)</ a >
2493
2493
</ li >
2494
2494
< li > < a href ="distributed.html#torch.distributed.init_process_group "> init_process_group() (in module torch.distributed)</ a >
2495
2495
</ li >
2496
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.init_rpc "> init_rpc() (in module torch.distributed.rpc)</ a >
2496
+ < li > < a href ="rpc.html#torch.distributed.rpc.init_rpc "> init_rpc() (in module torch.distributed.rpc)</ a >
2497
2497
</ li >
2498
2498
< li > < a href ="torch.html#torch.initial_seed "> initial_seed() (in module torch)</ a >
2499
2499
@@ -2603,7 +2603,7 @@ <h2 id="I">I</h2>
2603
2603
</ li >
2604
2604
< li > < a href ="distributed.html#torch.distributed.is_nccl_available "> is_nccl_available() (in module torch.distributed)</ a >
2605
2605
</ li >
2606
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.RRef.is_owner "> is_owner() (torch.distributed.rpc.RRef method)</ a >
2606
+ < li > < a href ="rpc.html#torch.distributed.rpc.RRef.is_owner "> is_owner() (torch.distributed.rpc.RRef method)</ a >
2607
2607
</ li >
2608
2608
< li > < a href ="storage.html#torch.FloatStorage.is_pinned "> is_pinned() (torch.FloatStorage method)</ a >
2609
2609
@@ -2877,7 +2877,7 @@ <h2 id="L">L</h2>
2877
2877
</ li >
2878
2878
< li > < a href ="nn.functional.html#torch.nn.functional.local_response_norm "> local_response_norm() (in module torch.nn.functional)</ a >
2879
2879
</ li >
2880
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.RRef.local_value "> local_value() (torch.distributed.rpc.RRef method)</ a >
2880
+ < li > < a href ="rpc.html#torch.distributed.rpc.RRef.local_value "> local_value() (torch.distributed.rpc.RRef method)</ a >
2881
2881
</ li >
2882
2882
< li > < a href ="nn.html#torch.nn.LocalResponseNorm "> LocalResponseNorm (class in torch.nn)</ a >
2883
2883
</ li >
@@ -3441,7 +3441,7 @@ <h2 id="M">M</h2>
3441
3441
< h2 id ="N "> N</ h2 >
3442
3442
< table style ="width: 100% " class ="indextable genindextable "> < tr >
3443
3443
< td style ="width: 33%; vertical-align: top; "> < ul >
3444
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.WorkerInfo.name "> name() (torch.distributed.rpc.WorkerInfo property)</ a >
3444
+ < li > < a href ="rpc.html#torch.distributed.rpc.WorkerInfo.name "> name() (torch.distributed.rpc.WorkerInfo property)</ a >
3445
3445
</ li >
3446
3446
< li > < a href ="nn.html#torch.nn.Module.named_buffers "> named_buffers() (torch.nn.Module method)</ a >
3447
3447
</ li >
@@ -3549,7 +3549,7 @@ <h2 id="N">N</h2>
3549
3549
< li > < a href ="torchvision/transforms.html#torchvision.transforms.functional.normalize "> (in module torchvision.transforms.functional)</ a >
3550
3550
</ li >
3551
3551
</ ul > </ li >
3552
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.ProcessGroupRpcBackendOptions.num_send_recv_threads "> num_send_recv_threads() (torch.distributed.rpc.ProcessGroupRpcBackendOptions property)</ a >
3552
+ < li > < a href ="rpc.html#torch.distributed.rpc.ProcessGroupRpcBackendOptions.num_send_recv_threads "> num_send_recv_threads() (torch.distributed.rpc.ProcessGroupRpcBackendOptions property)</ a >
3553
3553
</ li >
3554
3554
< li > < a href ="torch.html#torch.numel "> numel() (in module torch)</ a >
3555
3555
@@ -3605,9 +3605,9 @@ <h2 id="O">O</h2>
3605
3605
</ ul > </ li >
3606
3606
< li > < a href ="nn.init.html#torch.nn.init.orthogonal_ "> orthogonal_() (in module torch.nn.init)</ a >
3607
3607
</ li >
3608
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.RRef.owner "> owner() (torch.distributed.rpc.RRef method)</ a >
3608
+ < li > < a href ="rpc.html#torch.distributed.rpc.RRef.owner "> owner() (torch.distributed.rpc.RRef method)</ a >
3609
3609
</ li >
3610
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.RRef.owner_name "> owner_name() (torch.distributed.rpc.RRef method)</ a >
3610
+ < li > < a href ="rpc.html#torch.distributed.rpc.RRef.owner_name "> owner_name() (torch.distributed.rpc.RRef method)</ a >
3611
3611
</ li >
3612
3612
</ ul > </ td >
3613
3613
</ tr > </ table >
@@ -3775,7 +3775,7 @@ <h2 id="P">P</h2>
3775
3775
< li > < a href ="distributions.html#torch.distributions.relaxed_categorical.RelaxedOneHotCategorical.probs "> (torch.distributions.relaxed_categorical.RelaxedOneHotCategorical property)</ a >
3776
3776
</ li >
3777
3777
</ ul > </ li >
3778
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.ProcessGroupRpcBackendOptions "> ProcessGroupRpcBackendOptions (class in torch.distributed.rpc)</ a >
3778
+ < li > < a href ="rpc.html#torch.distributed.rpc.ProcessGroupRpcBackendOptions "> ProcessGroupRpcBackendOptions (class in torch.distributed.rpc)</ a >
3779
3779
</ li >
3780
3780
< li > < a href ="torch.html#torch.prod "> prod() (in module torch)</ a > , < a href ="torch.html#torch.prod "> [1]</ a > , < a href ="torch.html#torch.prod "> [2]</ a >
3781
3781
@@ -4033,7 +4033,7 @@ <h2 id="R">R</h2>
4033
4033
</ ul > </ li >
4034
4034
< li > < a href ="tensors.html#torch.Tensor.remainder_ "> remainder_() (torch.Tensor method)</ a >
4035
4035
</ li >
4036
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.remote "> remote() (in module torch.distributed.rpc)</ a >
4036
+ < li > < a href ="rpc.html#torch.distributed.rpc.remote "> remote() (in module torch.distributed.rpc)</ a >
4037
4037
</ li >
4038
4038
< li > < a href ="nn.html#torch.nn.utils.prune.remove "> remove() (in module torch.nn.utils.prune)</ a >
4039
4039
@@ -4187,15 +4187,15 @@ <h2 id="R">R</h2>
4187
4187
</ ul > </ li >
4188
4188
< li > < a href ="tensors.html#torch.Tensor.round_ "> round_() (torch.Tensor method)</ a >
4189
4189
</ li >
4190
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.rpc_async "> rpc_async() (in module torch.distributed.rpc)</ a >
4190
+ < li > < a href ="rpc.html#torch.distributed.rpc.rpc_async "> rpc_async() (in module torch.distributed.rpc)</ a >
4191
4191
</ li >
4192
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.rpc_sync "> rpc_sync() (in module torch.distributed.rpc)</ a >
4192
+ < li > < a href ="rpc.html#torch.distributed.rpc.rpc_sync "> rpc_sync() (in module torch.distributed.rpc)</ a >
4193
4193
</ li >
4194
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.ProcessGroupRpcBackendOptions.rpc_timeout "> rpc_timeout() (torch.distributed.rpc.ProcessGroupRpcBackendOptions property)</ a >
4194
+ < li > < a href ="rpc.html#torch.distributed.rpc.ProcessGroupRpcBackendOptions.rpc_timeout "> rpc_timeout() (torch.distributed.rpc.ProcessGroupRpcBackendOptions property)</ a >
4195
4195
</ li >
4196
4196
< li > < a href ="optim.html#torch.optim.Rprop "> Rprop (class in torch.optim)</ a >
4197
4197
</ li >
4198
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.RRef "> RRef (class in torch.distributed.rpc)</ a >
4198
+ < li > < a href ="rpc.html#torch.distributed.rpc.RRef "> RRef (class in torch.distributed.rpc)</ a >
4199
4199
</ li >
4200
4200
< li > < a href ="nn.html#torch.nn.RReLU "> RReLU (class in torch.nn)</ a >
4201
4201
</ li >
@@ -4449,7 +4449,7 @@ <h2 id="S">S</h2>
4449
4449
</ li >
4450
4450
< li > < a href ="torchvision/models.html#torchvision.models.shufflenet_v2_x2_0 "> shufflenet_v2_x2_0() (in module torchvision.models)</ a >
4451
4451
</ li >
4452
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.shutdown "> shutdown() (in module torch.distributed.rpc)</ a >
4452
+ < li > < a href ="rpc.html#torch.distributed.rpc.shutdown "> shutdown() (in module torch.distributed.rpc)</ a >
4453
4453
</ li >
4454
4454
< li > < a href ="nn.html#torch.nn.Sigmoid "> Sigmoid (class in torch.nn)</ a >
4455
4455
</ li >
@@ -4656,7 +4656,7 @@ <h2 id="S">S</h2>
4656
4656
< li > < a href ="amp.html#torch.cuda.amp.GradScaler.step "> step() (torch.cuda.amp.GradScaler method)</ a >
4657
4657
4658
4658
< ul >
4659
- < li > < a href ="rpc/rpc .html#torch.distributed.optim.DistributedOptimizer.step "> (torch.distributed.optim.DistributedOptimizer method)</ a >
4659
+ < li > < a href ="rpc.html#torch.distributed.optim.DistributedOptimizer.step "> (torch.distributed.optim.DistributedOptimizer method)</ a >
4660
4660
</ li >
4661
4661
< li > < a href ="optim.html#torch.optim.Adadelta.step "> (torch.optim.Adadelta method)</ a >
4662
4662
</ li >
@@ -4955,7 +4955,7 @@ <h2 id="T">T</h2>
4955
4955
</ li >
4956
4956
< li > < a href ="torchvision/transforms.html#torchvision.transforms.functional.to_grayscale "> to_grayscale() (in module torchvision.transforms.functional)</ a >
4957
4957
</ li >
4958
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.RRef.to_here "> to_here() (torch.distributed.rpc.RRef method)</ a >
4958
+ < li > < a href ="rpc.html#torch.distributed.rpc.RRef.to_here "> to_here() (torch.distributed.rpc.RRef method)</ a >
4959
4959
</ li >
4960
4960
< li > < a href ="tensors.html#torch.Tensor.to_mkldnn "> to_mkldnn() (torch.Tensor method)</ a >
4961
4961
</ li >
@@ -5011,13 +5011,13 @@ <h2 id="T">T</h2>
5011
5011
</ li >
5012
5012
< li > < a href ="distributed.html#module-torch.distributed "> torch.distributed (module)</ a >
5013
5013
</ li >
5014
- < li > < a href ="rpc/rpc .html#module-torch.distributed.autograd "> torch.distributed.autograd (module)</ a >
5014
+ < li > < a href ="rpc.html#module-torch.distributed.autograd "> torch.distributed.autograd (module)</ a >
5015
5015
</ li >
5016
5016
< li > < a href ="distributed.html#module-torch.distributed.launch "> torch.distributed.launch (module)</ a >
5017
5017
</ li >
5018
- < li > < a href ="rpc/rpc .html#module-torch.distributed.optim "> torch.distributed.optim (module)</ a >
5018
+ < li > < a href ="rpc.html#module-torch.distributed.optim "> torch.distributed.optim (module)</ a >
5019
5019
</ li >
5020
- < li > < a href ="rpc/rpc .html#module-torch.distributed.rpc "> torch.distributed.rpc (module)</ a >
5020
+ < li > < a href ="rpc.html#module-torch.distributed.rpc "> torch.distributed.rpc (module)</ a >
5021
5021
</ li >
5022
5022
< li > < a href ="distributions.html#module-torch.distributions "> torch.distributions (module)</ a >
5023
5023
</ li >
@@ -5449,7 +5449,7 @@ <h2 id="W">W</h2>
5449
5449
</ li >
5450
5450
< li > < a href ="torchvision/models.html#torchvision.models.wide_resnet50_2 "> wide_resnet50_2() (in module torchvision.models)</ a >
5451
5451
</ li >
5452
- < li > < a href ="rpc/rpc .html#torch.distributed.rpc.WorkerInfo "> WorkerInfo (class in torch.distributed.rpc)</ a >
5452
+ < li > < a href ="rpc.html#torch.distributed.rpc.WorkerInfo "> WorkerInfo (class in torch.distributed.rpc)</ a >
5453
5453
</ li >
5454
5454
< li > < a href ="torchvision/io.html#torchvision.io.write_video "> write_video() (in module torchvision.io)</ a >
5455
5455
</ li >
@@ -5782,4 +5782,4 @@ <h2>Resources</h2>
5782
5782
} )
5783
5783
</ script >
5784
5784
</ body >
5785
- </ html >
5785
+ </ html >
0 commit comments