We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 506c0ad commit 42c185aCopy full SHA for 42c185a
llama.cpp
@@ -4741,6 +4741,7 @@ static bool llm_load_tensors(
4741
model.buft_layer[i] = llama_default_buffer_type_cpu(true);
4742
}
4743
4744
+#ifndef GGML_USE_KOMPUTE
4745
if (split_mode == LLAMA_SPLIT_MODE_LAYER) {
4746
// calculate the split points
4747
int device_count = llama_get_device_count();
@@ -4778,7 +4779,9 @@ static bool llm_load_tensors(
4778
4779
} else {
4780
model.buft_output = llama_default_buffer_type_cpu(true);
4781
- } else {
4782
+ } else
4783
+#endif
4784
+ {
4785
ggml_backend_buffer_type_t split_buft;
4786
if (split_mode == LLAMA_SPLIT_MODE_ROW) {
4787
split_buft = llama_default_buffer_type_split(main_gpu, tensor_split);
0 commit comments