235
235
< div class ="pytorch-left-menu-search ">
236
236
237
237
< div class ="version ">
238
- < a href ='https://pytorch.org/docs/versions.html '> master (1.14.0a0+gitc77368d ) ▼</ a >
238
+ < a href ='https://pytorch.org/docs/versions.html '> master (1.14.0a0+git9f58e02 ) ▼</ a >
239
239
</ div >
240
240
241
241
@@ -580,7 +580,7 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
580
580
< span class ="s2 "> "different type."</ span >
581
581
< span class ="p "> )</ span >
582
582
< span class ="k "> else</ span > < span class ="p "> :</ span >
583
- < span class ="n "> new_storage</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage </ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> __deepcopy__ </ span > < span class ="p "> (</ span > < span class ="n "> memo</ span > < span class ="p "> )</ span >
583
+ < span class ="n "> new_storage</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> _typed_storage </ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> _deepcopy </ span > < span class ="p "> (</ span > < span class ="n "> memo</ span > < span class ="p "> )</ span >
584
584
< span class ="k "> if</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> is_quantized</ span > < span class ="p "> :</ span >
585
585
< span class ="c1 "> # quantizer_params can be different type based on torch attribute</ span >
586
586
< span class ="n "> quantizer_params</ span > < span class ="p "> :</ span > < span class ="n "> Union</ span > < span class ="p "> [</ span >
@@ -611,7 +611,9 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
611
611
< span class ="c1 "> # need to wrap with TypedStorage</ span >
612
612
< span class ="n "> new_tensor</ span > < span class ="o "> =</ span > < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> _utils</ span > < span class ="o "> .</ span > < span class ="n "> _rebuild_qtensor</ span > < span class ="p "> (</ span >
613
613
< span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="o "> .</ span > < span class ="n "> TypedStorage</ span > < span class ="p "> (</ span >
614
- < span class ="n "> wrap_storage</ span > < span class ="o "> =</ span > < span class ="n "> new_storage</ span > < span class ="o "> .</ span > < span class ="n "> untyped</ span > < span class ="p "> (),</ span > < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> dtype</ span >
614
+ < span class ="n "> wrap_storage</ span > < span class ="o "> =</ span > < span class ="n "> new_storage</ span > < span class ="o "> .</ span > < span class ="n "> _untyped_storage</ span > < span class ="p "> ,</ span >
615
+ < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> dtype</ span > < span class ="p "> ,</ span >
616
+ < span class ="n "> _internal</ span > < span class ="o "> =</ span > < span class ="kc "> True</ span > < span class ="p "> ,</ span >
615
617
< span class ="p "> ),</ span >
616
618
< span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage_offset</ span > < span class ="p "> (),</ span >
617
619
< span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> size</ span > < span class ="p "> (),</ span >
@@ -705,7 +707,17 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
705
707
< span class ="k "> if</ span > < span class ="n "> has_torch_function_unary</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
706
708
< span class ="k "> return</ span > < span class ="n "> handle_torch_function</ span > < span class ="p "> (</ span > < span class ="n "> Tensor</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="p "> ,</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,),</ span > < span class ="bp "> self</ span > < span class ="p "> )</ span >
707
709
708
- < span class ="k "> return</ span > < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> TypedStorage</ span > < span class ="p "> (</ span > < span class ="n "> wrap_storage</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> _storage</ span > < span class ="p "> (),</ span > < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> dtype</ span > < span class ="p "> )</ span > </ div >
710
+ < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="o "> .</ span > < span class ="n "> _warn_typed_storage_removal</ span > < span class ="p "> ()</ span >
711
+ < span class ="k "> return</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> _typed_storage</ span > < span class ="p "> ()</ span > </ div >
712
+
713
+ < span class ="c1 "> # For internal use only, to avoid raising deprecation warning</ span >
714
+ < span class ="k "> def</ span > < span class ="nf "> _typed_storage</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
715
+ < span class ="n "> _storage</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> _storage</ span > < span class ="p "> ()</ span >
716
+ < span class ="k "> if</ span > < span class ="nb "> isinstance</ span > < span class ="p "> (</ span > < span class ="n "> _storage</ span > < span class ="p "> ,</ span > < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> TypedStorage</ span > < span class ="p "> ):</ span >
717
+ < span class ="n "> _storage</ span > < span class ="o "> =</ span > < span class ="n "> _storage</ span > < span class ="o "> .</ span > < span class ="n "> _untyped_storage</ span >
718
+ < span class ="k "> return</ span > < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> TypedStorage</ span > < span class ="p "> (</ span >
719
+ < span class ="n "> wrap_storage</ span > < span class ="o "> =</ span > < span class ="n "> _storage</ span > < span class ="p "> ,</ span > < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> dtype</ span > < span class ="p "> ,</ span > < span class ="n "> _internal</ span > < span class ="o "> =</ span > < span class ="kc "> True</ span >
720
+ < span class ="p "> )</ span >
709
721
710
722
< span class ="k "> def</ span > < span class ="nf "> _reduce_ex_internal</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="n "> proto</ span > < span class ="p "> ):</ span >
711
723
< span class ="n "> check_serializing_named_tensor</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> )</ span >
@@ -779,7 +791,9 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
779
791
< span class ="c1 "> # need to wrap with TypedStorage</ span >
780
792
< span class ="n "> args_qtensor</ span > < span class ="o "> =</ span > < span class ="p "> (</ span >
781
793
< span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="o "> .</ span > < span class ="n "> TypedStorage</ span > < span class ="p "> (</ span >
782
- < span class ="n "> wrap_storage</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> untyped</ span > < span class ="p "> (),</ span > < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> dtype</ span >
794
+ < span class ="n "> wrap_storage</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> _typed_storage</ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> _untyped_storage</ span > < span class ="p "> ,</ span >
795
+ < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> dtype</ span > < span class ="p "> ,</ span >
796
+ < span class ="n "> _internal</ span > < span class ="o "> =</ span > < span class ="kc "> True</ span > < span class ="p "> ,</ span >
783
797
< span class ="p "> ),</ span >
784
798
< span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage_offset</ span > < span class ="p "> (),</ span >
785
799
< span class ="nb "> tuple</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> size</ span > < span class ="p "> ()),</ span >
@@ -837,7 +851,9 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
837
851
< span class ="c1 "> # need to wrap with TypedStorage</ span >
838
852
< span class ="n "> args</ span > < span class ="o "> =</ span > < span class ="p "> (</ span >
839
853
< span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="o "> .</ span > < span class ="n "> TypedStorage</ span > < span class ="p "> (</ span >
840
- < span class ="n "> wrap_storage</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> untyped</ span > < span class ="p "> (),</ span > < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> dtype</ span >
854
+ < span class ="n "> wrap_storage</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> _typed_storage</ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> _untyped_storage</ span > < span class ="p "> ,</ span >
855
+ < span class ="n "> dtype</ span > < span class ="o "> =</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> dtype</ span > < span class ="p "> ,</ span >
856
+ < span class ="n "> _internal</ span > < span class ="o "> =</ span > < span class ="kc "> True</ span > < span class ="p "> ,</ span >
841
857
< span class ="p "> ),</ span >
842
858
< span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage_offset</ span > < span class ="p "> (),</ span >
843
859
< span class ="nb "> tuple</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> size</ span > < span class ="p "> ()),</ span >
@@ -1055,7 +1071,7 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
1055
1071
< span class ="sd "> """</ span >
1056
1072
< span class ="k "> if</ span > < span class ="n "> has_torch_function_unary</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
1057
1073
< span class ="k "> return</ span > < span class ="n "> handle_torch_function</ span > < span class ="p "> (</ span > < span class ="n "> Tensor</ span > < span class ="o "> .</ span > < span class ="n "> is_shared</ span > < span class ="p "> ,</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,),</ span > < span class ="bp "> self</ span > < span class ="p "> )</ span >
1058
- < span class ="k "> return</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage </ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> is_shared </ span > < span class ="p "> ()</ span > </ div >
1074
+ < span class ="k "> return</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> _typed_storage </ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> _is_shared </ span > < span class ="p "> ()</ span > </ div >
1059
1075
1060
1076
< div class ="viewcode-block " id ="Tensor.share_memory_ "> < a class ="viewcode-back " href ="../../generated/torch.Tensor.share_memory_.html#torch.Tensor.share_memory_ "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> share_memory_</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
1061
1077
< span class ="sa "> r</ span > < span class ="sd "> """Moves the underlying storage to shared memory.</ span >
@@ -1065,7 +1081,7 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
1065
1081
< span class ="sd "> """</ span >
1066
1082
< span class ="k "> if</ span > < span class ="n "> has_torch_function_unary</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
1067
1083
< span class ="k "> return</ span > < span class ="n "> handle_torch_function</ span > < span class ="p "> (</ span > < span class ="n "> Tensor</ span > < span class ="o "> .</ span > < span class ="n "> share_memory_</ span > < span class ="p "> ,</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,),</ span > < span class ="bp "> self</ span > < span class ="p "> )</ span >
1068
- < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage </ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> share_memory_ </ span > < span class ="p "> ()</ span >
1084
+ < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> _typed_storage </ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> _share_memory_ </ span > < span class ="p "> ()</ span >
1069
1085
< span class ="k "> return</ span > < span class ="bp "> self</ span > </ div >
1070
1086
1071
1087
< span class ="k "> def</ span > < span class ="fm "> __reversed__</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
@@ -1507,7 +1523,9 @@ <h1>Source code for torch._tensor</h1><div class="highlight"><pre>
1507
1523
< span class ="k "> if</ span > < span class ="n "> has_torch_function_unary</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ):</ span >
1508
1524
< span class ="k "> return</ span > < span class ="n "> handle_torch_function</ span > < span class ="p "> (</ span > < span class ="n "> Tensor</ span > < span class ="o "> .</ span > < span class ="n "> storage_type</ span > < span class ="p "> ,</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,),</ span > < span class ="bp "> self</ span > < span class ="p "> )</ span >
1509
1525
1510
- < span class ="k "> return</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> _get_legacy_storage_class</ span > < span class ="p "> ()</ span > </ div >
1526
+ < span class ="n "> torch</ span > < span class ="o "> .</ span > < span class ="n "> storage</ span > < span class ="o "> .</ span > < span class ="n "> _warn_typed_storage_removal</ span > < span class ="p "> ()</ span >
1527
+
1528
+ < span class ="k "> return</ span > < span class ="bp "> self</ span > < span class ="o "> .</ span > < span class ="n "> _typed_storage</ span > < span class ="p "> ()</ span > < span class ="o "> .</ span > < span class ="n "> _get_legacy_storage_class</ span > < span class ="p "> ()</ span > </ div >
1511
1529
1512
1530
< div class ="viewcode-block " id ="Tensor.refine_names "> < a class ="viewcode-back " href ="../../named_tensor.html#torch.Tensor.refine_names "> [docs]</ a > < span class ="k "> def</ span > < span class ="nf "> refine_names</ span > < span class ="p "> (</ span > < span class ="bp "> self</ span > < span class ="p "> ,</ span > < span class ="o "> *</ span > < span class ="n "> names</ span > < span class ="p "> ):</ span >
1513
1531
< span class ="sa "> r</ span > < span class ="sd "> """Refines the dimension names of :attr:`self` according to :attr:`names`.</ span >
0 commit comments