We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4a06b01 commit dbf17f0Copy full SHA for dbf17f0
llama.cpp
@@ -3763,6 +3763,7 @@ static bool llm_load_tensors(
3763
model.buft_layer[i] = llama_default_buffer_type_cpu(true);
3764
}
3765
3766
+#ifndef GGML_USE_KOMPUTE
3767
if (split_mode == LLAMA_SPLIT_MODE_LAYER) {
3768
// calculate the split points
3769
int device_count = llama_get_device_count();
@@ -3800,7 +3801,9 @@ static bool llm_load_tensors(
3800
3801
} else {
3802
model.buft_output = llama_default_buffer_type_cpu(true);
3803
- } else {
3804
+ } else
3805
+#endif
3806
+ {
3807
ggml_backend_buffer_type_t split_buft;
3808
if (split_mode == LLAMA_SPLIT_MODE_ROW) {
3809
split_buft = llama_default_buffer_type_split(main_gpu, tensor_split);
0 commit comments