File tree 1 file changed +8
-0
lines changed 1 file changed +8
-0
lines changed Original file line number Diff line number Diff line change @@ -3346,7 +3346,11 @@ static size_t llama_get_device_count(const llama_model & model) {
3346
3346
static ggml_backend_buffer_type_t llama_default_buffer_type_offload(const llama_model & model, int gpu) {
3347
3347
ggml_backend_buffer_type_t buft = nullptr;
3348
3348
3349
+ #ifdef GGML_USE_RPC
3349
3350
int rpc_count = (int)model.rpc_servers.size();
3351
+ #else
3352
+ int rpc_count = 0;
3353
+ #endif
3350
3354
int local_gpu = gpu - rpc_count;
3351
3355
#if defined(GGML_USE_RPC)
3352
3356
if (gpu < rpc_count) {
@@ -3403,7 +3407,11 @@ static ggml_backend_buffer_type_t llama_default_buffer_type_split(const llama_mo
3403
3407
}
3404
3408
3405
3409
static size_t llama_get_device_memory(const llama_model & model, int device) {
3410
+ #ifdef GGML_USE_RPC
3406
3411
int rpc_count = (int)model.rpc_servers.size();
3412
+ #else
3413
+ int rpc_count = 0;
3414
+ #endif
3407
3415
int local_device = device - rpc_count;
3408
3416
#if defined(GGML_USE_RPC)
3409
3417
if (device < rpc_count) {
You can’t perform that action at this time.
0 commit comments