Skip to content

Commit b5b9e4b

Browse files
committed
fix(convert): Remove unused tensor name mappings
Branch: GraniteMoE Co-Authored-By: [email protected] Signed-off-by: Gabe Goodhart <[email protected]>
1 parent 1349625 commit b5b9e4b

File tree

1 file changed

+8
-10
lines changed

1 file changed

+8
-10
lines changed

gguf-py/gguf/tensor_mapping.py

Lines changed: 8 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -293,11 +293,10 @@ class TensorNameMap:
293293
),
294294

295295
MODEL_TENSOR.FFN_UP_EXP: (
296-
"layers.{bid}.feed_forward.experts.w3", # mixtral (merged)
297-
"transformer.decoder_layer.{bid}.moe.linear_v", # Grok (merged)
298-
"transformer.blocks.{bid}.ffn.experts.mlp.v1", # dbrx
299-
"model.layers.{bid}.mlp.experts.up_proj", # qwen2moe olmoe (merged)
300-
"model.layers.{bid}.block_sparse_moe.input_linear.up", # granitemoe
296+
"layers.{bid}.feed_forward.experts.w3", # mixtral (merged)
297+
"transformer.decoder_layer.{bid}.moe.linear_v", # Grok (merged)
298+
"transformer.blocks.{bid}.ffn.experts.mlp.v1", # dbrx
299+
"model.layers.{bid}.mlp.experts.up_proj", # qwen2moe olmoe (merged)
301300
),
302301

303302
MODEL_TENSOR.FFN_UP_SHEXP: (
@@ -326,11 +325,10 @@ class TensorNameMap:
326325
),
327326

328327
MODEL_TENSOR.FFN_GATE_EXP: (
329-
"layers.{bid}.feed_forward.experts.w1", # mixtral (merged)
330-
"transformer.decoder_layer.{bid}.moe.linear", # Grok (merged)
331-
"transformer.blocks.{bid}.ffn.experts.mlp.w1", # dbrx
332-
"model.layers.{bid}.mlp.experts.gate_proj", # qwen2moe olmoe (merged)
333-
"model.layers.{bid}.block_sparse_moe.input_linear.gate", # granitemoe
328+
"layers.{bid}.feed_forward.experts.w1", # mixtral (merged)
329+
"transformer.decoder_layer.{bid}.moe.linear", # Grok (merged)
330+
"transformer.blocks.{bid}.ffn.experts.mlp.w1", # dbrx
331+
"model.layers.{bid}.mlp.experts.gate_proj", # qwen2moe olmoe (merged)
334332
),
335333

336334
MODEL_TENSOR.FFN_GATE_SHEXP: (

0 commit comments

Comments
 (0)