We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent f273c23 commit df9c5f3Copy full SHA for df9c5f3
src/llama.cpp
@@ -6872,6 +6872,7 @@ static bool llm_load_tensors(
6872
model.buft_layer[i] = llama_default_buffer_type_cpu(true);
6873
}
6874
6875
+#ifndef GGML_USE_KOMPUTE
6876
if (split_mode == LLAMA_SPLIT_MODE_LAYER) {
6877
// calculate the split points
6878
int device_count = llama_get_device_count(model);
@@ -6909,7 +6910,9 @@ static bool llm_load_tensors(
6909
6910
} else {
6911
model.buft_output = llama_default_buffer_type_cpu(true);
6912
- } else {
6913
+ } else
6914
+#endif
6915
+ {
6916
ggml_backend_buffer_type_t split_buft;
6917
if (split_mode == LLAMA_SPLIT_MODE_ROW) {
6918
split_buft = llama_default_buffer_type_split(model, main_gpu, tensor_split);
0 commit comments