@@ -2673,40 +2673,40 @@ static llama_control_vector_data llama_control_vector_load_one(const llama_contr
2673
2673
uint32_t layer = std::stoi (name.substr (dotpos + 1 ));
2674
2674
if (layer == 0 ) {
2675
2675
fprintf (stderr, " %s: direction tensor invalid in %s\n " , __func__, load_info.fname .c_str ());
2676
- ggml_free (meta_ctx);
2677
2676
gguf_free (meta_ctx_gguf);
2677
+ ggml_free (meta_ctx);
2678
2678
return result;
2679
2679
}
2680
2680
if (layer > max_direction_layer) {
2681
2681
max_direction_layer = layer;
2682
2682
}
2683
2683
} catch (...) {
2684
2684
fprintf (stderr, " %s: direction tensor invalid in %s\n " , __func__, load_info.fname .c_str ());
2685
- ggml_free (meta_ctx);
2686
2685
gguf_free (meta_ctx_gguf);
2686
+ ggml_free (meta_ctx);
2687
2687
return result;
2688
2688
}
2689
2689
}
2690
2690
2691
2691
struct ggml_tensor * tensor_meta = ggml_get_tensor (meta_ctx, name.c_str ());
2692
2692
if (tensor_meta->type != GGML_TYPE_F32 || ggml_n_dims (tensor_meta) != 1 ) {
2693
2693
fprintf (stderr, " %s: direction tensor invalid in %s\n " , __func__, load_info.fname .c_str ());
2694
- ggml_free (meta_ctx);
2695
2694
gguf_free (meta_ctx_gguf);
2695
+ ggml_free (meta_ctx);
2696
2696
return result;
2697
2697
}
2698
2698
if (result.n_embd == -1 ) {
2699
2699
result.n_embd = ggml_nelements (tensor_meta);
2700
2700
} else if (ggml_nelements (tensor_meta) != result.n_embd ) {
2701
2701
fprintf (stderr, " %s: direction tensor sizes mismatched in %s\n " , __func__, load_info.fname .c_str ());
2702
- ggml_free (meta_ctx);
2703
2702
gguf_free (meta_ctx_gguf);
2703
+ ggml_free (meta_ctx);
2704
2704
return result;
2705
2705
}
2706
2706
n_bytes += ggml_nbytes (tensor_meta);
2707
2707
}
2708
- ggml_free (meta_ctx);
2709
2708
gguf_free (meta_ctx_gguf);
2709
+ ggml_free (meta_ctx);
2710
2710
}
2711
2711
2712
2712
if (n_tensors == 0 ) {
0 commit comments