Skip to content

Commit c082b9f

Browse files
authored
llama.cpp : use rope scale kv
1 parent dc1f051 commit c082b9f

File tree

1 file changed

+8
-0
lines changed

1 file changed

+8
-0
lines changed

llama.cpp

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1435,6 +1435,14 @@ static void llama_model_load_internal(
14351435
hparams.n_head_kv = hparams.n_head;
14361436
GGUF_GET(hparams.n_head_kv, gguf_get_val_u32, GGUF_TYPE_UINT32, false, "llama.attention.head_count_kv");
14371437

1438+
// TODO: manually setting rope scale should override this
1439+
// rope_freq_scale (inverse of the kv) is optional
1440+
float ropescale = 1.0f;
1441+
GGUF_GET(ropescale, gguf_get_val_f32, GGUF_TYPE_FLOAT32, false, "llama.rope.scale_linear");
1442+
if (ropescale != 1.0f) {
1443+
rope_freq_scale = 1.0f/ropescale;
1444+
}
1445+
14381446
// get general kv
14391447
GGUF_GET(general_name, gguf_get_val_str, GGUF_TYPE_STRING, false, "general.name");
14401448
GGUF_GET(general_arch, gguf_get_val_str, GGUF_TYPE_STRING, false, "general.architecture");

0 commit comments

Comments
 (0)