Skip to content

Commit 56eb18b

Browse files
authored
Pipe in local_global attention (#10883)
Differential Revision: D74762916
1 parent 879235b commit 56eb18b

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

examples/models/llama/export_llama_lib.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -719,6 +719,7 @@ def _prepare_for_llama_export(args) -> LLMEdgeManager:
719719
preq_mode=args.preq_mode,
720720
preq_group_size=args.preq_group_size,
721721
preq_embedding_quantize=args.preq_embedding_quantize,
722+
local_global_attention=args.local_global_attention,
722723
)
723724
)
724725

@@ -1447,7 +1448,7 @@ def _get_source_transforms( # noqa
14471448
transforms.append(
14481449
partial(
14491450
replace_kv_cache_with_ring_kv_cache,
1450-
layer_sizes=args.local_global_attention,
1451+
layer_sizes=local_global_attention,
14511452
)
14521453
)
14531454

0 commit comments

Comments
 (0)