Skip to content

Commit 2a1ace5

Browse files
committed
Update on "allow customized head_dim"
This is for resolving the ask in this [post](https://fb.workplace.com/groups/pytorch.edge.users/permalink/1574875706716050/). Similar change in HF: huggingface/transformers#32502 Differential Revision: [D65974454](https://our.internmc.facebook.com/intern/diff/D65974454/) [ghstack-poisoned]
2 parents 1c87ce3 + fa35c88 commit 2a1ace5

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

examples/models/llama/llama_transformer.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -85,7 +85,7 @@ class ModelArgs:
8585
n_kv_heads: Optional[int] = None
8686
vocab_size: int = -1 # defined later by tokenizer
8787
hidden_dim: Optional[int] = None
88-
head_dim: Optional[int] = None
88+
head_dim: Optional[int] = None # Optional customized head_dim
8989
multiple_of: int = 256 # make SwiGLU hidden layer size multiple of large power of 2
9090
ffn_dim_multiplier: Optional[float] = None
9191
norm_eps: float = 1e-5

0 commit comments

Comments
 (0)