Skip to content

Commit f34ffbc

Browse files
committed
llama: fix inference with RWKV6Qwen2
Signed-off-by: Molly Sophia <[email protected]>
1 parent ba7bdc0 commit f34ffbc

File tree

1 file changed

+0
-4
lines changed

1 file changed

+0
-4
lines changed

src/llama-model.cpp

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -10648,8 +10648,6 @@ struct llm_build_rwkv6qwen2 : public llm_build_rwkv6_base {
1064810648
const auto n_seq_tokens = ubatch.n_seq_tokens;
1064910649
const auto n_seqs = ubatch.n_seqs;
1065010650

10651-
inpL = build_inp_embd(model.tok_embd);
10652-
1065310651
for (int il = 0; il < n_layer; ++il) {
1065410652
const llama_layer * layer = &model.layers[il];
1065510653

@@ -11001,8 +10999,6 @@ struct llm_build_arwkv7 : public llm_build_rwkv7_base {
1100110999
const auto n_seq_tokens = ubatch.n_seq_tokens;
1100211000
const auto n_seqs = ubatch.n_seqs;
1100311001

11004-
inpL = build_inp_embd(model.tok_embd);
11005-
1100611002
for (int il = 0; il < n_layer; ++il) {
1100711003
const llama_layer * layer = &model.layers[il];
1100811004

0 commit comments

Comments
 (0)