Skip to content

Commit 271560e

Browse files
committed
Don't test for DEEPSEEK2 arch only as others might use MLA in future
1 parent 1df57db commit 271560e

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

src/llama-kv-cache.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@ llama_kv_cache_unified::llama_kv_cache_unified(
3232
uint32_t padding) : model(model), hparams(model.hparams), v_trans(v_trans), padding(padding) {
3333
const int32_t n_layer = hparams.n_layer;
3434

35-
const bool is_mla = model.arch == LLM_ARCH_DEEPSEEK2 && hparams.n_embd_head_k_mla != 0 && hparams.n_embd_head_v_mla != 0;
35+
const bool is_mla = (hparams.n_embd_head_k_mla != 0 && hparams.n_embd_head_v_mla != 0);
3636

3737
has_shift = false;
3838
can_shift = !is_mla || v_trans; // TODO: allow context shifting for MLA with flash attention

0 commit comments

Comments
 (0)