@@ -322,7 +322,7 @@ def ggml_pad(x: int, n: int) -> int:
322
322
323
323
def add_tensor_info (
324
324
self , name : str , tensor_shape : Sequence [int ], tensor_dtype : np .dtype ,
325
- tensor_nbytes : int , raw_dtype : GGMLQuantizationType | None = None ,
325
+ tensor_nbytes : int , raw_dtype : int | None = None ,
326
326
) -> None :
327
327
if self .state is not WriterState .NO_FILE :
328
328
raise ValueError (f'Expected output file to be not yet opened, got { self .state } ' )
@@ -348,7 +348,7 @@ def add_tensor_info(
348
348
else :
349
349
raise ValueError ("Only F16, F32, F64, I8, I16, I32, I64 tensors are supported for now" )
350
350
else :
351
- dtype = raw_dtype
351
+ dtype = GGMLQuantizationType ( raw_dtype )
352
352
if tensor_dtype == np .uint8 :
353
353
tensor_shape = quant_shape_from_byte_shape (tensor_shape , raw_dtype )
354
354
@@ -367,7 +367,7 @@ def add_tensor_info(
367
367
368
368
def add_tensor (
369
369
self , name : str , tensor : np .ndarray [Any , Any ], raw_shape : Sequence [int ] | None = None ,
370
- raw_dtype : GGMLQuantizationType | None = None ,
370
+ raw_dtype : int | None = None ,
371
371
) -> None :
372
372
if self .endianess == GGUFEndian .BIG :
373
373
tensor .byteswap (inplace = True )
0 commit comments