207
207
< div class ="pytorch-left-menu-search ">
208
208
209
209
< div class ="version ">
210
- < a href ='https://pytorch.org/docs/versions.html '> master (1.11.0a0+git8fa5cde ) ▼</ a >
210
+ < a href ='https://pytorch.org/docs/versions.html '> master (1.11.0a0+gitcf70466 ) ▼</ a >
211
211
</ div >
212
212
213
213
@@ -702,7 +702,7 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
702
702
< span class ="c1 "> # All strings are unicode in Python 3.</ span >
703
703
< span class ="k "> return</ span > < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> _tensor_str</ span > < span class ="o "> .</ span > < span class ="n "> _str</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> )</ span >
704
704
705
- < span class ="k "> def</ span > < span class ="nf "> backward</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> gradient</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ,</ span > < span class ="n "> retain_graph</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ,</ span > < span class ="n "> create_graph</ span > < span class ="o "> =</ span > < span class ="kc "> False</ span > < span class ="p "> ,</ span > < span class ="n "> inputs</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ):</ span >
705
+ < div class =" viewcode-block " id =" Tensor.backward " > < a class =" viewcode-back " href =" ../../generated/torch.Tensor.backward.html#torch.Tensor.backward " > [docs] </ a > < span class ="k "> def</ span > < span class ="nf "> backward</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> gradient</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ,</ span > < span class ="n "> retain_graph</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ,</ span > < span class ="n "> create_graph</ span > < span class ="o "> =</ span > < span class ="kc "> False</ span > < span class ="p "> ,</ span > < span class ="n "> inputs</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ):</ span >
706
706
< span class ="sa "> r</ span > < span class ="sd "> """Computes the gradient of current tensor w.r.t. graph leaves.</ span >
707
707
708
708
< span class ="sd "> The graph is differentiated using the chain rule. If the tensor is</ span >
@@ -758,7 +758,7 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
758
758
< span class ="n "> retain_graph</ span > < span class ="o "> =</ span > < span class ="n "> retain_graph</ span > < span class ="p "> ,</ span >
759
759
< span class ="n "> create_graph</ span > < span class ="o "> =</ span > < span class ="n "> create_graph</ span > < span class ="p "> ,</ span >
760
760
< span class ="n "> inputs</ span > < span class ="o "> =</ span > < span class ="n "> inputs</ span > < span class ="p "> )</ span >
761
- < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> autograd</ span > < span class ="o "> .</ span > < span class ="n "> backward</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> gradient</ span > < span class ="p "> ,</ span > < span class ="n "> retain_graph</ span > < span class ="p "> ,</ span > < span class ="n "> create_graph</ span > < span class ="p "> ,</ span > < span class ="n "> inputs</ span > < span class ="o "> =</ span > < span class ="n "> inputs</ span > < span class ="p "> )</ span >
761
+ < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> autograd</ span > < span class ="o "> .</ span > < span class ="n "> backward</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> gradient</ span > < span class ="p "> ,</ span > < span class ="n "> retain_graph</ span > < span class ="p "> ,</ span > < span class ="n "> create_graph</ span > < span class ="p "> ,</ span > < span class ="n "> inputs</ span > < span class ="o "> =</ span > < span class ="n "> inputs</ span > < span class ="p "> )</ span > </ div >
762
762
763
763
< span class ="k "> def</ span > < span class ="nf "> register_hook</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> hook</ span > < span class ="p "> ):</ span >
764
764
< span class ="sa "> r</ span > < span class ="sd "> """Registers a backward hook.</ span >
@@ -895,7 +895,7 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
895
895
< span class ="k "> return</ span > < span class ="n "> handle_torch_function</ span > < span class ="p "> (</ span > < span class ="n "> Tensor</ span > < span class ="o "> .</ span > < span class ="n "> norm</ span > < span class ="p "> ,</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,),</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> p</ span > < span class ="o "> =</ span > < span class ="n "> p</ span > < span class ="p "> ,</ span > < span class ="n "> dim</ span > < span class ="o "> =</ span > < span class ="n "> dim</ span > < span class ="p "> ,</ span > < span class ="n "> keepdim</ span > < span class ="o "> =</ span > < span class ="n "> keepdim</ span > < span class ="p "> ,</ span > < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="n "> dtype</ span > < span class ="p "> )</ span >
896
896
< span class ="k "> return</ span > < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> norm</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> p</ span > < span class ="p "> ,</ span > < span class ="n "> dim</ span > < span class ="p "> ,</ span > < span class ="n "> keepdim</ span > < span class ="p "> ,</ span > < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="n "> dtype</ span > < span class ="p "> )</ span > </ div >
897
897
898
- < div class =" viewcode-block " id =" Tensor.lu " > < a class =" viewcode-back " href =" ../../generated/torch.Tensor.lu.html#torch.Tensor.lu " > [docs] </ a > < span class ="k "> def</ span > < span class ="nf "> lu</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> pivot</ span > < span class ="o "> =</ span > < span class ="kc "> True</ span > < span class ="p "> ,</ span > < span class ="n "> get_infos</ span > < span class ="o "> =</ span > < span class ="kc "> False</ span > < span class ="p "> ):</ span >
898
+ < span class ="k "> def</ span > < span class ="nf "> lu</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> pivot</ span > < span class ="o "> =</ span > < span class ="kc "> True</ span > < span class ="p "> ,</ span > < span class ="n "> get_infos</ span > < span class ="o "> =</ span > < span class ="kc "> False</ span > < span class ="p "> ):</ span >
899
899
< span class ="sa "> r</ span > < span class ="sd "> """See :func:`torch.lu`"""</ span >
900
900
< span class ="c1 "> # If get_infos is True, then we don't need to check for errors and vice versa</ span >
901
901
< span class ="k "> if</ span > < span class ="n "> has_torch_function_unary</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
@@ -905,7 +905,7 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
905
905
< span class ="k "> if</ span > < span class ="n "> get_infos</ span > < span class ="p "> :</ span >
906
906
< span class ="k "> return</ span > < span class ="n "> LU</ span > < span class ="p "> ,</ span > < span class ="n "> pivots</ span > < span class ="p "> ,</ span > < span class ="n "> infos</ span >
907
907
< span class ="k "> else</ span > < span class ="p "> :</ span >
908
- < span class ="k "> return</ span > < span class ="n "> LU</ span > < span class ="p "> ,</ span > < span class ="n "> pivots</ span > </ div >
908
+ < span class ="k "> return</ span > < span class ="n "> LU</ span > < span class ="p "> ,</ span > < span class ="n "> pivots</ span >
909
909
910
910
< span class ="k "> def</ span > < span class ="nf "> stft</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> n_fft</ span > < span class ="p "> :</ span > < span class ="nb "> int</ span > < span class ="p "> ,</ span > < span class ="n "> hop_length</ span > < span class ="p "> :</ span > < span class ="n "> Optional</ span > < span class ="p "> [</ span > < span class ="nb "> int</ span > < span class ="p "> ]</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ,</ span >
911
911
< span class ="n "> win_length</ span > < span class ="p "> :</ span > < span class ="n "> Optional</ span > < span class ="p "> [</ span > < span class ="nb "> int</ span > < span class ="p "> ]</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ,</ span > < span class ="n "> window</ span > < span class ="p "> :</ span > < span class ="s1 "> 'Optional[Tensor]'</ span > < span class ="o "> =</ span > < span class ="kc "> None</ span > < span class ="p "> ,</ span >
0 commit comments