@@ -15391,23 +15391,45 @@ static ggml_type llama_tensor_get_type(quantize_state_internal & qs, ggml_type n
15391
15391
else if (ftype == LLAMA_FTYPE_MOSTLY_Q4_K_S && qs.i_attention_wv < 4) new_type = GGML_TYPE_Q5_K;
15392
15392
++qs.i_attention_wv;
15393
15393
} else if (name.find("attn_k.weight") != std::string::npos) {
15394
- if (ftype == LLAMA_FTYPE_MOSTLY_IQ2_XXS || ftype == LLAMA_FTYPE_MOSTLY_IQ2_XS || ftype == LLAMA_FTYPE_MOSTLY_IQ1_S ||
15395
- ftype == LLAMA_FTYPE_MOSTLY_IQ2_S || ftype == LLAMA_FTYPE_MOSTLY_IQ1_M) {
15396
- if (qs.model.hparams.n_expert >= 8) new_type = GGML_TYPE_Q5_K;
15397
- }
15398
- else if (qs.model.hparams.n_expert >= 8) {
15394
+ if (qs.model.hparams.n_expert >= 8) {
15399
15395
// for the 8-expert model, bumping this to Q8_0 trades just ~128MB
15400
15396
// TODO: explore better strategies
15401
- new_type = GGML_TYPE_Q8_0;
15397
+ if (ftype == LLAMA_FTYPE_MOSTLY_IQ2_XXS || ftype == LLAMA_FTYPE_MOSTLY_IQ2_XS || ftype == LLAMA_FTYPE_MOSTLY_IQ1_S ||
15398
+ ftype == LLAMA_FTYPE_MOSTLY_IQ2_S || ftype == LLAMA_FTYPE_MOSTLY_IQ1_M || ftype == LLAMA_FTYPE_MOSTLY_Q2_K_S ||
15399
+ ftype == LLAMA_FTYPE_MOSTLY_Q2_K) new_type = GGML_TYPE_Q6_K;
15400
+ else new_type = GGML_TYPE_Q8_0;
15402
15401
}
15403
- else if (ftype == LLAMA_FTYPE_MOSTLY_IQ3_M && (qs.model.hparams.n_gqa() >= 2 || qs.model.hparams.n_expert >= 2)) {
15404
- new_type = GGML_TYPE_IQ4_XS;
15402
+ else if ((ftype == LLAMA_FTYPE_MOSTLY_IQ1_S) && (qs.model.hparams.n_gqa() >= 2 || qs.model.hparams.n_expert >= 2)) {
15403
+ new_type = GGML_TYPE_IQ1_M;
15404
+ }
15405
+ else if ((ftype == LLAMA_FTYPE_MOSTLY_IQ1_M) && (qs.model.hparams.n_gqa() >= 2 || qs.model.hparams.n_expert >= 2)) {
15406
+ new_type = GGML_TYPE_IQ2_XXS;
15407
+ }
15408
+ else if ((ftype == LLAMA_FTYPE_MOSTLY_IQ2_XXS) && (qs.model.hparams.n_gqa() >= 2 || qs.model.hparams.n_expert >= 2)) {
15409
+ new_type = GGML_TYPE_IQ2_XS;
15410
+ }
15411
+ else if ((ftype == LLAMA_FTYPE_MOSTLY_IQ2_XS) && (qs.model.hparams.n_gqa() >= 2 || qs.model.hparams.n_expert >= 2)) {
15412
+ new_type = GGML_TYPE_IQ2_S;
15413
+ }
15414
+ else if ((ftype == LLAMA_FTYPE_MOSTLY_IQ2_S || ftype == LLAMA_FTYPE_MOSTLY_IQ2_M) &&
15415
+ (qs.model.hparams.n_gqa() >= 2 || qs.model.hparams.n_expert >= 2)) {
15416
+ new_type = GGML_TYPE_IQ3_XXS;
15417
+ }
15418
+ else if (ftype == LLAMA_FTYPE_MOSTLY_Q2_K && (qs.model.hparams.n_gqa() >= 2 || qs.model.hparams.n_expert >= 2)) {
15419
+ new_type = GGML_TYPE_Q3_K;
15405
15420
}
15406
15421
else if (ftype == LLAMA_FTYPE_MOSTLY_IQ3_XS && qs.model.hparams.n_gqa() < 2 && qs.model.hparams.n_expert < 2) {
15407
15422
new_type = GGML_TYPE_IQ3_XXS;
15408
15423
}
15409
- else if (ftype == LLAMA_FTYPE_MOSTLY_IQ3_XXS && qs.model.hparams.n_gqa() < 2 && qs.model.hparams.n_expert < 2) {
15410
- new_type = GGML_TYPE_IQ2_S;
15424
+ else if (ftype == LLAMA_FTYPE_MOSTLY_IQ3_M && (qs.model.hparams.n_gqa() >= 2 || qs.model.hparams.n_expert >= 2)) {
15425
+ new_type = GGML_TYPE_IQ4_XS;
15426
+ }
15427
+ else if (ftype == LLAMA_FTYPE_MOSTLY_Q3_K_L && (qs.model.hparams.n_gqa() >= 2 || qs.model.hparams.n_expert >= 2)) {
15428
+ new_type = GGML_TYPE_Q4_K;
15429
+ }
15430
+ else if ((ftype == LLAMA_FTYPE_MOSTLY_IQ4_XS || ftype == LLAMA_FTYPE_MOSTLY_IQ4_NL || ftype == LLAMA_FTYPE_MOSTLY_Q4_K_S ||
15431
+ ftype == LLAMA_FTYPE_MOSTLY_Q4_K_M) && (qs.model.hparams.n_gqa() >= 2 || qs.model.hparams.n_expert >= 2)) {
15432
+ new_type = GGML_TYPE_Q5_K;
15411
15433
}
15412
15434
} else if (name.find("attn_q.weight") != std::string::npos) {
15413
15435
if (ftype == LLAMA_FTYPE_MOSTLY_IQ3_XS) new_type = GGML_TYPE_IQ3_XXS;
0 commit comments