Skip to content

Commit 62642bb

Browse files
authored
talk-llama : fix build after ggml sync (ggml-org#1049)
sed -i 's,GGML_BACKEND_CUDA,GGML_BACKEND_GPU,g' examples/talk-llama/llama.cpp
1 parent f1c9df5 commit 62642bb

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

examples/talk-llama/llama.cpp

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1002,7 +1002,7 @@ static void llama_model_load_internal(
10021002
}
10031003

10041004
#ifdef GGML_USE_CUBLAS
1005-
#define LLAMA_BACKEND_OFFLOAD GGML_BACKEND_CUDA
1005+
#define LLAMA_BACKEND_OFFLOAD GGML_BACKEND_GPU
10061006
#else
10071007
#define LLAMA_BACKEND_OFFLOAD GGML_BACKEND_CPU
10081008
#endif
@@ -1054,7 +1054,7 @@ static void llama_model_load_internal(
10541054
layer.w2 = ml->get_tensor(layers_i + ".feed_forward.w2.weight", { n_ff, n_embd}, backend);
10551055
layer.w3 = ml->get_tensor(layers_i + ".feed_forward.w3.weight", {n_embd, n_ff}, backend);
10561056

1057-
if (backend == GGML_BACKEND_CUDA) {
1057+
if (backend == GGML_BACKEND_GPU) {
10581058
vram_total +=
10591059
ggml_nbytes(layer.attention_norm) + ggml_nbytes(layer.wq) + ggml_nbytes(layer.wk) +
10601060
ggml_nbytes(layer.wv) + ggml_nbytes(layer.wo) + ggml_nbytes(layer.attention_norm) +
@@ -1115,7 +1115,7 @@ static void llama_model_load_internal(
11151115
}
11161116
}
11171117
for (llama_load_tensor & lt : ml->tensors_map.tensors) {
1118-
if (lt.ggml_tensor->backend != GGML_BACKEND_CUDA) {
1118+
if (lt.ggml_tensor->backend != GGML_BACKEND_GPU) {
11191119
continue;
11201120
}
11211121
if (progress_callback) {

0 commit comments

Comments
 (0)