We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 879235b commit 56eb18bCopy full SHA for 56eb18b
examples/models/llama/export_llama_lib.py
@@ -719,6 +719,7 @@ def _prepare_for_llama_export(args) -> LLMEdgeManager:
719
preq_mode=args.preq_mode,
720
preq_group_size=args.preq_group_size,
721
preq_embedding_quantize=args.preq_embedding_quantize,
722
+ local_global_attention=args.local_global_attention,
723
)
724
725
@@ -1447,7 +1448,7 @@ def _get_source_transforms( # noqa
1447
1448
transforms.append(
1449
partial(
1450
replace_kv_cache_with_ring_kv_cache,
- layer_sizes=args.local_global_attention,
1451
+ layer_sizes=local_global_attention,
1452
1453
1454
0 commit comments