Skip to content

Commit 4713a40

Browse files
committed
LLAMA_BACKEND_OFFLOAD* -> llama_backend_offload*
1 parent 698c945 commit 4713a40

File tree

1 file changed

+5
-5
lines changed

1 file changed

+5
-5
lines changed

llama.cpp

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -3130,12 +3130,12 @@ static void llm_load_tensors(
31303130
// norm is not performance relevant on its own but keeping it in VRAM reduces data copying
31313131
// on Windows however this is detrimental unless everything is on the GPU
31323132
#ifndef _WIN32
3133-
backend_norm = LLAMA_BACKEND_OFFLOAD;
3133+
backend_norm = llama_backend_offload;
31343134
#else
3135-
backend_norm = n_gpu_layers <= (int) n_layer + 2 ? GGML_BACKEND_CPU : LLAMA_BACKEND_OFFLOAD;
3135+
backend_norm = n_gpu_layers <= (int) n_layer + 2 ? GGML_BACKEND_CPU : llama_backend_offload;
31363136
#endif // _WIN32
31373137

3138-
backend_output = LLAMA_BACKEND_OFFLOAD_SPLIT;
3138+
backend_output = llama_backend_offload_split;
31393139
} else {
31403140
backend_norm = GGML_BACKEND_CPU;
31413141
backend_output = GGML_BACKEND_CPU;
@@ -3163,8 +3163,8 @@ static void llm_load_tensors(
31633163
/*
31643164
llama_model_loader: - tensor 4: blk.0.attn_output.weight f16 [ 2560, 2560, 1, 1 ]
31653165
*/
3166-
const ggml_backend_type backend = int(i) < i_gpu_start ? GGML_BACKEND_CPU : LLAMA_BACKEND_OFFLOAD; // NOLINT
3167-
const ggml_backend_type backend_split = int(i) < i_gpu_start ? GGML_BACKEND_CPU : LLAMA_BACKEND_OFFLOAD_SPLIT; // NOLINT
3166+
const ggml_backend_type backend = int(i) < i_gpu_start ? GGML_BACKEND_CPU : llama_backend_offload; // NOLINT
3167+
const ggml_backend_type backend_split = int(i) < i_gpu_start ? GGML_BACKEND_CPU : llama_backend_offload_split; // NOLINT
31683168

31693169
auto & layer = model.layers[i];
31703170

0 commit comments

Comments
 (0)