@@ -1377,7 +1377,6 @@ def main(args_in: list[str] | None = None) -> None:
1377
1377
# We currently only support Q8_0 output on little endian systems.
1378
1378
output_choices .append ("q8_0" )
1379
1379
parser = argparse .ArgumentParser (description = "Convert a LLaMA model to a GGML compatible file" )
1380
- parser .add_argument ("--awq-path" , type = Path , help = "Path to scale awq cache file" , default = None )
1381
1380
parser .add_argument ("--dump" , action = "store_true" , help = "don't convert, just show what's in the model" )
1382
1381
parser .add_argument ("--dump-single" , action = "store_true" , help = "don't convert, just show what's in a single model file" )
1383
1382
parser .add_argument ("--vocab-only" , action = "store_true" , help = "extract only the vocab" )
@@ -1393,18 +1392,6 @@ def main(args_in: list[str] | None = None) -> None:
1393
1392
parser .add_argument ("--skip-unknown" , action = "store_true" , help = "skip unknown tensor names instead of failing" )
1394
1393
1395
1394
args = parser .parse_args (args_in )
1396
- if args .awq_path :
1397
- sys .path .insert (1 , str (Path (__file__ ).parent / 'awq-py' ))
1398
- from awq .apply_awq import add_scale_weights # type: ignore[import-not-found]
1399
- tmp_model_path = args .model / "weighted_model"
1400
- if tmp_model_path .is_dir ():
1401
- print (f"{ tmp_model_path } exists as a weighted model." )
1402
- else :
1403
- tmp_model_path .mkdir (parents = True , exist_ok = True )
1404
- print ("Saving new weighted model ..." )
1405
- add_scale_weights (str (args .model ), str (args .awq_path ), str (tmp_model_path ))
1406
- print (f"Saved weighted model at { tmp_model_path } ." )
1407
- args .model = tmp_model_path
1408
1395
1409
1396
if args .dump_single :
1410
1397
model_plus = lazy_load_file (args .model )
0 commit comments