230
230
< div class ="pytorch-left-menu-search ">
231
231
232
232
< div class ="version ">
233
- < a href ='https://pytorch.org/docs/versions.html '> main (2.1.0a0+git5ebb18c ) ▼</ a >
233
+ < a href ='https://pytorch.org/docs/versions.html '> main (2.1.0a0+gitb3c29cd ) ▼</ a >
234
234
</ div >
235
235
236
236
@@ -618,7 +618,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
618
618
619
619
620
620
< span class ="c1 "> # See Note [Global dependencies]</ span >
621
- < span class ="k "> def</ span > < span class ="nf "> _load_global_deps</ span > < span class ="p "> ():</ span >
621
+ < span class ="k "> def</ span > < span class ="nf "> _load_global_deps</ span > < span class ="p "> ()</ span > < span class =" o " > -> </ span > < span class =" kc " > None </ span > < span class =" p " > :</ span >
622
622
< span class ="k "> if</ span > < span class ="n "> _running_with_deploy</ span > < span class ="p "> ()</ span > < span class ="ow "> or</ span > < span class ="n "> platform</ span > < span class ="o "> .</ span > < span class ="n "> system</ span > < span class ="p "> ()</ span > < span class ="o "> ==</ span > < span class ="s1 "> 'Windows'</ span > < span class ="p "> :</ span >
623
623
< span class ="k "> return</ span >
624
624
@@ -1116,7 +1116,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
1116
1116
< span class ="sd "> """</ span >
1117
1117
< span class ="n "> _C</ span > < span class ="o "> .</ span > < span class ="n "> _set_default_dtype</ span > < span class ="p "> (</ span > < span class ="n "> d</ span > < span class ="p "> )</ span > </ div >
1118
1118
1119
- < div class ="viewcode-block " id ="use_deterministic_algorithms "> < a class ="viewcode-back " href ="../generated/torch.use_deterministic_algorithms.html#torch.use_deterministic_algorithms "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> use_deterministic_algorithms</ span > < span class ="p "> (</ span > < span class ="n "> mode</ span > < span class ="p "> ,</ span > < span class ="o "> *</ span > < span class ="p "> ,</ span > < span class ="n "> warn_only</ span > < span class ="o "> = </ span > < span class ="kc "> False</ span > < span class ="p "> ):</ span >
1119
+ < div class ="viewcode-block " id ="use_deterministic_algorithms "> < a class ="viewcode-back " href ="../generated/torch.use_deterministic_algorithms.html#torch.use_deterministic_algorithms "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> use_deterministic_algorithms</ span > < span class ="p "> (</ span > < span class ="n "> mode</ span > < span class ="p "> : </ span > < span class =" n " > builtins </ span > < span class =" o " > . </ span > < span class =" n " > bool </ span > < span class =" p " > ,</ span > < span class ="o "> *</ span > < span class ="p "> ,</ span > < span class ="n "> warn_only</ span > < span class ="p " > : </ span > < span class =" n " > builtins </ span > < span class =" o "> . </ span > < span class =" n " > bool </ span > < span class ="o " > = </ span > < span class =" kc "> False</ span > < span class ="p "> )</ span > < span class =" o " > -> </ span > < span class =" kc " > None </ span > < span class =" p " > :</ span >
1120
1120
< span class ="w "> </ span > < span class ="sa "> r</ span > < span class ="sd "> """ Sets whether PyTorch operations must use "deterministic"</ span >
1121
1121
< span class ="sd "> algorithms. That is, algorithms which, given the same input, and when</ span >
1122
1122
< span class ="sd "> run on the same software and hardware, always produce the same output.</ span >
@@ -1256,13 +1256,13 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
1256
1256
< span class ="sd "> """</ span >
1257
1257
< span class ="n "> _C</ span > < span class ="o "> .</ span > < span class ="n "> _set_deterministic_algorithms</ span > < span class ="p "> (</ span > < span class ="n "> mode</ span > < span class ="p "> ,</ span > < span class ="n "> warn_only</ span > < span class ="o "> =</ span > < span class ="n "> warn_only</ span > < span class ="p "> )</ span > </ div >
1258
1258
1259
- < div class ="viewcode-block " id ="are_deterministic_algorithms_enabled "> < a class ="viewcode-back " href ="../generated/torch.are_deterministic_algorithms_enabled.html#torch.are_deterministic_algorithms_enabled "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> are_deterministic_algorithms_enabled</ span > < span class ="p "> ():</ span >
1259
+ < div class ="viewcode-block " id ="are_deterministic_algorithms_enabled "> < a class ="viewcode-back " href ="../generated/torch.are_deterministic_algorithms_enabled.html#torch.are_deterministic_algorithms_enabled "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> are_deterministic_algorithms_enabled</ span > < span class ="p "> ()</ span > < span class =" o " > -> </ span > < span class =" n " > builtins </ span > < span class =" o " > . </ span > < span class =" n " > bool </ span > < span class =" p " > :</ span >
1260
1260
< span class ="w "> </ span > < span class ="sa "> r</ span > < span class ="sd "> """Returns True if the global deterministic flag is turned on. Refer to</ span >
1261
1261
< span class ="sd "> :func:`torch.use_deterministic_algorithms` documentation for more details.</ span >
1262
1262
< span class ="sd "> """</ span >
1263
1263
< span class ="k "> return</ span > < span class ="n "> _C</ span > < span class ="o "> .</ span > < span class ="n "> _get_deterministic_algorithms</ span > < span class ="p "> ()</ span > </ div >
1264
1264
1265
- < div class ="viewcode-block " id ="is_deterministic_algorithms_warn_only_enabled "> < a class ="viewcode-back " href ="../generated/torch.is_deterministic_algorithms_warn_only_enabled.html#torch.is_deterministic_algorithms_warn_only_enabled "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> is_deterministic_algorithms_warn_only_enabled</ span > < span class ="p "> ():</ span >
1265
+ < div class ="viewcode-block " id ="is_deterministic_algorithms_warn_only_enabled "> < a class ="viewcode-back " href ="../generated/torch.is_deterministic_algorithms_warn_only_enabled.html#torch.is_deterministic_algorithms_warn_only_enabled "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> is_deterministic_algorithms_warn_only_enabled</ span > < span class ="p "> ()</ span > < span class =" o " > -> </ span > < span class =" n " > builtins </ span > < span class =" o " > . </ span > < span class =" n " > bool </ span > < span class =" p " > :</ span >
1266
1266
< span class ="w "> </ span > < span class ="sa "> r</ span > < span class ="sd "> """Returns True if the global deterministic flag is set to warn only.</ span >
1267
1267
< span class ="sd "> Refer to :func:`torch.use_deterministic_algorithms` documentation for more</ span >
1268
1268
< span class ="sd "> details.</ span >
@@ -1331,7 +1331,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
1331
1331
< span class ="sd "> """</ span >
1332
1332
< span class ="k "> return</ span > < span class ="n "> _C</ span > < span class ="o "> .</ span > < span class ="n "> _get_float32_matmul_precision</ span > < span class ="p "> ()</ span > </ div >
1333
1333
1334
- < div class ="viewcode-block " id ="set_float32_matmul_precision "> < a class ="viewcode-back " href ="../generated/torch.set_float32_matmul_precision.html#torch.set_float32_matmul_precision "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> set_float32_matmul_precision</ span > < span class ="p "> (</ span > < span class ="n "> precision</ span > < span class ="p "> ) :</ span >
1334
+ < div class ="viewcode-block " id ="set_float32_matmul_precision "> < a class ="viewcode-back " href ="../generated/torch.set_float32_matmul_precision.html#torch.set_float32_matmul_precision "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> set_float32_matmul_precision</ span > < span class ="p "> (</ span > < span class ="n "> precision</ span > < span class ="p "> : </ span > < span class =" nb " > str </ span > < span class =" p " > ) </ span > < span class =" o " > -> </ span > < span class =" kc " > None </ span > < span class =" p " > :</ span >
1335
1335
< span class ="w "> </ span > < span class ="sa "> r</ span > < span class ="sd "> """Sets the internal precision of float32 matrix multiplications.</ span >
1336
1336
1337
1337
< span class ="sd "> Running float32 matrix multiplications in lower precision may significantly increase</ span >
@@ -1376,7 +1376,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
1376
1376
< span class ="sd "> """</ span >
1377
1377
< span class ="n "> _C</ span > < span class ="o "> .</ span > < span class ="n "> _set_float32_matmul_precision</ span > < span class ="p "> (</ span > < span class ="n "> precision</ span > < span class ="p "> )</ span > </ div >
1378
1378
1379
- < div class ="viewcode-block " id ="set_warn_always "> < a class ="viewcode-back " href ="../generated/torch.set_warn_always.html#torch.set_warn_always "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> set_warn_always</ span > < span class ="p "> (</ span > < span class ="n "> b</ span > < span class ="p "> ) :</ span >
1379
+ < div class ="viewcode-block " id ="set_warn_always "> < a class ="viewcode-back " href ="../generated/torch.set_warn_always.html#torch.set_warn_always "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> set_warn_always</ span > < span class ="p "> (</ span > < span class ="n "> b</ span > < span class ="p "> : </ span > < span class =" n " > builtins </ span > < span class =" o " > . </ span > < span class =" n " > bool </ span > < span class =" p " > ) </ span > < span class =" o " > -> </ span > < span class =" kc " > None </ span > < span class =" p " > :</ span >
1380
1380
< span class ="w "> </ span > < span class ="sa "> r</ span > < span class ="sd "> """When this flag is False (default) then some PyTorch warnings may only</ span >
1381
1381
< span class ="sd "> appear once per process. This helps avoid excessive warning information.</ span >
1382
1382
< span class ="sd "> Setting it to True causes these warnings to always appear, which may be</ span >
@@ -1388,7 +1388,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
1388
1388
< span class ="sd "> """</ span >
1389
1389
< span class ="n "> _C</ span > < span class ="o "> .</ span > < span class ="n "> _set_warnAlways</ span > < span class ="p "> (</ span > < span class ="n "> b</ span > < span class ="p "> )</ span > </ div >
1390
1390
1391
- < div class ="viewcode-block " id ="is_warn_always_enabled "> < a class ="viewcode-back " href ="../generated/torch.is_warn_always_enabled.html#torch.is_warn_always_enabled "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> is_warn_always_enabled</ span > < span class ="p "> ():</ span >
1391
+ < div class ="viewcode-block " id ="is_warn_always_enabled "> < a class ="viewcode-back " href ="../generated/torch.is_warn_always_enabled.html#torch.is_warn_always_enabled "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> is_warn_always_enabled</ span > < span class ="p "> ()</ span > < span class =" o " > -> </ span > < span class =" n " > builtins </ span > < span class =" o " > . </ span > < span class =" n " > bool </ span > < span class =" p " > :</ span >
1392
1392
< span class ="w "> </ span > < span class ="sa "> r</ span > < span class ="sd "> """Returns True if the global warn_always flag is turned on. Refer to</ span >
1393
1393
< span class ="sd "> :func:`torch.set_warn_always` documentation for more details.</ span >
1394
1394
< span class ="sd "> """</ span >
@@ -1896,7 +1896,7 @@ <h1>Source code for torch</h1><div class="highlight"><pre>
1896
1896
< span class ="k "> del</ span > < span class ="n "> _torch_docs</ span > < span class ="p "> ,</ span > < span class ="n "> _tensor_docs</ span > < span class ="p "> ,</ span > < span class ="n "> _storage_docs</ span >
1897
1897
1898
1898
1899
- < div class ="viewcode-block " id ="compiled_with_cxx11_abi "> < a class ="viewcode-back " href ="../generated/torch.compiled_with_cxx11_abi.html#torch.compiled_with_cxx11_abi "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> compiled_with_cxx11_abi</ span > < span class ="p "> ():</ span >
1899
+ < div class ="viewcode-block " id ="compiled_with_cxx11_abi "> < a class ="viewcode-back " href ="../generated/torch.compiled_with_cxx11_abi.html#torch.compiled_with_cxx11_abi "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> compiled_with_cxx11_abi</ span > < span class ="p "> ()</ span > < span class =" o " > -> </ span > < span class =" n " > builtins </ span > < span class =" o " > . </ span > < span class =" n " > bool </ span > < span class =" p " > :</ span >
1900
1900
< span class ="w "> </ span > < span class ="sa "> r</ span > < span class ="sd "> """Returns whether PyTorch was built with _GLIBCXX_USE_CXX11_ABI=1"""</ span >
1901
1901
< span class ="k "> return</ span > < span class ="n "> _C</ span > < span class ="o "> .</ span > < span class ="n "> _GLIBCXX_USE_CXX11_ABI</ span > </ div >
1902
1902
0 commit comments