Skip to content

Commit b73e564

Browse files
authored
quantize : fix precedence of cli args (#6541)
1 parent e3c337d commit b73e564

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

llama.cpp

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13562,10 +13562,10 @@ static void llama_model_quantize_internal(const std::string & fname_inp, const s
1356213562
if (!params->pure && ggml_is_quantized(default_type)) {
1356313563
new_type = llama_tensor_get_type(qs, new_type, tensor, ftype);
1356413564
}
13565-
else if (params->token_embedding_type < GGML_TYPE_COUNT && strcmp(tensor->name, "token_embd.weight") == 0) {
13565+
if (params->token_embedding_type < GGML_TYPE_COUNT && strcmp(tensor->name, "token_embd.weight") == 0) {
1356613566
new_type = params->token_embedding_type;
1356713567
}
13568-
else if (params->output_tensor_type < GGML_TYPE_COUNT && strcmp(tensor->name, "output.weight") == 0) {
13568+
if (params->output_tensor_type < GGML_TYPE_COUNT && strcmp(tensor->name, "output.weight") == 0) {
1356913569
new_type = params->output_tensor_type;
1357013570
}
1357113571

0 commit comments

Comments
 (0)