Skip to content

Commit 04d7e65

Browse files
committed
kompute : disable GPU offload for Mixtral
We haven't implemented the necessary GPU kernels yet. Fixes this crash: ggml_vk_graph_compute: error: unsupported op 'ARGSORT' GGML_ASSERT: /home/jared/src/forks/gpt4all/gpt4all-backend/llama.cpp-mainline/ggml-kompute.cpp:1508: !"unsupported op"
1 parent 9954f11 commit 04d7e65

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

src/llama.cpp

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9050,6 +9050,7 @@ static int llama_model_load(const std::string & fname, llama_model & model, llam
90509050
model.using_gpu = false;
90519051
} else if (
90529052
!(model.arch == LLM_ARCH_LLAMA || model.arch == LLM_ARCH_FALCON)
9053+
|| model.hparams.n_expert > 0
90539054
|| !(
90549055
model.ftype == LLAMA_FTYPE_ALL_F32 ||
90559056
model.ftype == LLAMA_FTYPE_MOSTLY_F16 ||

0 commit comments

Comments
 (0)