Skip to content

Commit f7260c2

Browse files
committed
qwen2.5vl ok
1 parent 792387b commit f7260c2

File tree

2 files changed

+6
-5
lines changed

2 files changed

+6
-5
lines changed

convert_hf_to_gguf.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1103,7 +1103,7 @@ def set_gguf_parameters(self):
11031103

11041104
# preprocessor config
11051105
self.gguf_writer.add_vision_image_mean(self.preprocessor_config["image_mean"])
1106-
self.gguf_writer.add_vision_image_std(self.preprocessor_config["image_mean"])
1106+
self.gguf_writer.add_vision_image_std(self.preprocessor_config["image_std"])
11071107

11081108
def write_vocab(self):
11091109
raise ValueError("VisionModel does not support vocab writing")
@@ -2563,8 +2563,9 @@ def __init__(self, *args, **kwargs):
25632563
# rename config.json values
25642564
self.hparams["num_attention_heads"] = self.hparams.get("num_heads")
25652565
self.hparams["num_hidden_layers"] = self.hparams.get("depth")
2566-
self.hparams["intermediate_size"] = self.hparams.get("hidden_size")
2567-
self.hparams["hidden_size"] = self.hparams.get("embed_dim")
2566+
if "embed_dim" in self.hparams: # qwen2vl
2567+
self.hparams["intermediate_size"] = self.hparams.get("hidden_size")
2568+
self.hparams["hidden_size"] = self.hparams.get("embed_dim")
25682569

25692570
def set_gguf_parameters(self):
25702571
super().set_gguf_parameters()

gguf-py/gguf/tensor_mapping.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -990,7 +990,7 @@ class TensorNameMap:
990990
"model.vision_model.encoder.layers.{bid}.mlp.fc2", # SmolVLM, gemma3 (note: name is swapped)
991991
"vision_tower.transformer.layers.{bid}.feed_forward.up_proj", # pixtral
992992
"visual.blocks.{bid}.mlp.fc2", # qwen2vl
993-
"visual.blocks.{bid}.mlp.down_proj", # qwen2.5vl
993+
"visual.blocks.{bid}.mlp.up_proj", # qwen2.5vl
994994
),
995995

996996
MODEL_TENSOR.V_ENC_FFN_GATE: (
@@ -1004,7 +1004,7 @@ class TensorNameMap:
10041004
"model.vision_model.encoder.layers.{bid}.mlp.fc1", # SmolVLM, gemma3 (note: name is swapped)
10051005
"vision_tower.transformer.layers.{bid}.feed_forward.down_proj", # pixtral
10061006
"visual.blocks.{bid}.mlp.fc1", # qwen2vl
1007-
"visual.blocks.{bid}.mlp.up_proj", # qwen2.5vl
1007+
"visual.blocks.{bid}.mlp.down_proj", # qwen2.5vl
10081008
),
10091009

10101010
MODEL_TENSOR.V_PRE_NORM: (

0 commit comments

Comments
 (0)