38
38
< link rel ="search " title ="Search " href ="../../search.html " />
39
39
40
40
41
+ <!-- Google Analytics -->
42
+
43
+ < script async src ="https://www.googletagmanager.com/gtag/js?id=UA-117752657-2 "> </ script >
44
+ < script >
45
+ window . dataLayer = window . dataLayer || [ ] ;
46
+ function gtag ( ) { dataLayer . push ( arguments ) ; }
47
+ gtag ( 'js' , new Date ( ) ) ;
48
+
49
+ gtag ( 'config' , 'UA-117752657-2' ) ;
50
+ </ script >
51
+
52
+ <!-- End Google Analytics -->
53
+
54
+
41
55
42
56
43
57
< script src ="../../_static/js/modernizr.min.js "> </ script >
@@ -784,14 +798,14 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
784
798
< span class ="s2 "> have forward mode AD gradients.</ span >
785
799
< span class ="s2 "> """</ span > < span class ="p "> )</ span >
786
800
787
- < span class ="k "> def</ span > < span class ="nf "> is_shared</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
801
+ < div class =" viewcode-block " id =" Tensor.is_shared " > < a class =" viewcode-back " href =" ../../generated/torch.Tensor.is_shared.html#torch.Tensor.is_shared " > [docs] </ a > < span class ="k "> def</ span > < span class ="nf "> is_shared</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
788
802
< span class ="sa "> r</ span > < span class ="sd "> """Checks if tensor is in shared memory.</ span >
789
803
790
804
< span class ="sd "> This is always ``True`` for CUDA tensors.</ span >
791
805
< span class ="sd "> """</ span >
792
806
< span class ="k "> if</ span > < span class ="n "> has_torch_function_unary</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
793
807
< span class ="k "> return</ span > < span class ="n "> handle_torch_function</ span > < span class ="p "> (</ span > < span class ="n "> Tensor</ span > < span class ="o "> .</ span > < span class ="n "> is_shared</ span > < span class ="p "> ,</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,),</ span > < span class ="bp "> self</ span > < span class ="p "> )</ span >
794
- < span class ="k "> return</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> is_shared</ span > < span class ="p "> ()</ span >
808
+ < span class ="k "> return</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> is_shared</ span > < span class ="p "> ()</ span > </ div >
795
809
796
810
< span class ="k "> def</ span > < span class ="nf "> share_memory_</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
797
811
< span class ="sa "> r</ span > < span class ="sd "> """Moves the underlying storage to shared memory.</ span >
@@ -819,7 +833,7 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
819
833
< span class ="k "> return</ span > < span class ="n "> handle_torch_function</ span > < span class ="p "> (</ span > < span class ="n "> Tensor</ span > < span class ="o "> .</ span > < span class ="n "> norm</ span > < span class ="p "> ,</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,),</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> p</ span > < span class ="o "> =</ span > < span class ="n "> p</ span > < span class ="p "> ,</ span > < span class ="n "> dim</ span > < span class ="o "> =</ span > < span class ="n "> dim</ span > < span class ="p "> ,</ span > < span class ="n "> keepdim</ span > < span class ="o "> =</ span > < span class ="n "> keepdim</ span > < span class ="p "> ,</ span > < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="n "> dtype</ span > < span class ="p "> )</ span >
820
834
< span class ="k "> return</ span > < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> norm</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> p</ span > < span class ="p "> ,</ span > < span class ="n "> dim</ span > < span class ="p "> ,</ span > < span class ="n "> keepdim</ span > < span class ="p "> ,</ span > < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="n "> dtype</ span > < span class ="p "> )</ span >
821
835
822
- < div class =" viewcode-block " id =" Tensor.lu " > < a class =" viewcode-back " href =" ../../generated/torch.Tensor.lu.html#torch.Tensor.lu " > [docs] </ a > < span class ="k "> def</ span > < span class ="nf "> lu</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> pivot</ span > < span class ="o "> =</ span > < span class ="kc "> True</ span > < span class ="p "> ,</ span > < span class ="n "> get_infos</ span > < span class ="o "> =</ span > < span class ="kc "> False</ span > < span class ="p "> ):</ span >
836
+ < span class ="k "> def</ span > < span class ="nf "> lu</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> pivot</ span > < span class ="o "> =</ span > < span class ="kc "> True</ span > < span class ="p "> ,</ span > < span class ="n "> get_infos</ span > < span class ="o "> =</ span > < span class ="kc "> False</ span > < span class ="p "> ):</ span >
823
837
< span class ="sa "> r</ span > < span class ="sd "> """See :func:`torch.lu`"""</ span >
824
838
< span class ="c1 "> # If get_infos is True, then we don't need to check for errors and vice versa</ span >
825
839
< span class ="k "> if</ span > < span class ="n "> has_torch_function_unary</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
@@ -829,7 +843,7 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
829
843
< span class ="k "> if</ span > < span class ="n "> get_infos</ span > < span class ="p "> :</ span >
830
844
< span class ="k "> return</ span > < span class ="n "> LU</ span > < span class ="p "> ,</ span > < span class ="n "> pivots</ span > < span class ="p "> ,</ span > < span class ="n "> infos</ span >
831
845
< span class ="k "> else</ span > < span class ="p "> :</ span >
832
- < span class ="k "> return</ span > < span class ="n "> LU</ span > < span class ="p "> ,</ span > < span class ="n "> pivots</ span > </ div >
846
+ < span class ="k "> return</ span > < span class ="n "> LU</ span > < span class ="p "> ,</ span > < span class ="n "> pivots</ span >
833
847
834
848
< span class ="k "> def</ span > < span class ="nf "> stft</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> n_fft</ span > < span class ="p "> :</ span > < span class ="nb "> int</ span > < span class ="p "> ,</ span > < span class ="n "> hop_length</ span > < span class ="p "> :</ span > < span class ="n "> Optional</ span > < span class ="p "> [</ span > < span class ="nb "> int</ span > < span class ="p "> ]</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ,</ span >
835
849
< span class ="n "> win_length</ span > < span class ="p "> :</ span > < span class ="n "> Optional</ span > < span class ="p "> [</ span > < span class ="nb "> int</ span > < span class ="p "> ]</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ,</ span > < span class ="n "> window</ span > < span class ="p "> :</ span > < span class ="s1 "> 'Optional[Tensor]'</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ,</ span >
0 commit comments