Skip to content

Commit df9c5f3

Browse files
committed
kompute : disable LLAMA_SPLIT_LAYER after ggml-org#5321
1 parent f273c23 commit df9c5f3

File tree

1 file changed

+4
-1
lines changed

1 file changed

+4
-1
lines changed

src/llama.cpp

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6872,6 +6872,7 @@ static bool llm_load_tensors(
68726872
model.buft_layer[i] = llama_default_buffer_type_cpu(true);
68736873
}
68746874

6875+
#ifndef GGML_USE_KOMPUTE
68756876
if (split_mode == LLAMA_SPLIT_MODE_LAYER) {
68766877
// calculate the split points
68776878
int device_count = llama_get_device_count(model);
@@ -6909,7 +6910,9 @@ static bool llm_load_tensors(
69096910
} else {
69106911
model.buft_output = llama_default_buffer_type_cpu(true);
69116912
}
6912-
} else {
6913+
} else
6914+
#endif
6915+
{
69136916
ggml_backend_buffer_type_t split_buft;
69146917
if (split_mode == LLAMA_SPLIT_MODE_ROW) {
69156918
split_buft = llama_default_buffer_type_split(model, main_gpu, tensor_split);

0 commit comments

Comments
 (0)