Skip to content

Commit 764b856

Browse files
CISCngxson
andauthored
convert : qwen2/3moe : set yarn metadata if present (#13331)
* set yarn metadata if present * add comment about enabling YaRN Co-authored-by: Xuan-Son Nguyen <[email protected]> --------- Co-authored-by: Xuan-Son Nguyen <[email protected]>
1 parent 15a28ec commit 764b856

File tree

1 file changed

+7
-0
lines changed

1 file changed

+7
-0
lines changed

convert_hf_to_gguf.py

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2761,6 +2761,13 @@ def set_gguf_parameters(self):
27612761
if (shared_expert_intermediate_size := self.hparams.get('shared_expert_intermediate_size')) is not None:
27622762
self.gguf_writer.add_expert_shared_feed_forward_length(shared_expert_intermediate_size)
27632763
logger.info(f"gguf: expert shared feed forward length = {shared_expert_intermediate_size}")
2764+
# YaRN is not enabled by default
2765+
# To enable it, please refer to this guide: https://huggingface.co/Qwen/Qwen3-30B-A3B#processing-long-texts
2766+
if self.hparams.get("rope_scaling") is not None and "factor" in self.hparams["rope_scaling"]:
2767+
if self.hparams["rope_scaling"].get("type") == "yarn":
2768+
self.gguf_writer.add_rope_scaling_type(gguf.RopeScalingType.YARN)
2769+
self.gguf_writer.add_rope_scaling_factor(self.hparams["rope_scaling"]["factor"])
2770+
self.gguf_writer.add_rope_scaling_orig_ctx_len(self.hparams["rope_scaling"]["original_max_position_embeddings"])
27642771

27652772
_experts: list[dict[str, Tensor]] | None = None
27662773

0 commit comments

Comments
 (0)