Skip to content

Commit fa2c0d5

Browse files
committed
finetune: rename feed-forward tensors (w1/w2/w3)
This commit renames the feed-forward tensors w1, w2 and w3 to ffn_gate, ffn_down and ffn_up respectively. The motivation for this change is to make it easier to understand the purpose of the tensors. This also seems to be inline with the names used in the llama_layer struct in llama.cpp. Signed-off-by: Daniel Bevenius <[email protected]>
1 parent 49cc1f7 commit fa2c0d5

File tree

2 files changed

+124
-124
lines changed

2 files changed

+124
-124
lines changed

examples/finetune/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -80,9 +80,9 @@ The LORA rank can be configured for each model tensor type separately with these
8080
--rank-wk N LORA rank for wk tensor (default 4)
8181
--rank-wv N LORA rank for wv tensor (default 4)
8282
--rank-wo N LORA rank for wo tensor (default 4)
83-
--rank-w1 N LORA rank for w1 tensor (default 4)
84-
--rank-w2 N LORA rank for w2 tensor (default 4)
85-
--rank-w3 N LORA rank for w3 tensor (default 4)
83+
--rank-ffn_gate N LORA rank for ffn_gate tensor (default 4)
84+
--rank-ffn_down N LORA rank for ffn_down tensor (default 4)
85+
--rank-ffn_up N LORA rank for ffn_up tensor (default 4)
8686
```
8787

8888
The LORA rank of 'norm' tensors should always be 1.

0 commit comments

Comments
 (0)