Skip to content

Commit 96ba40b

Browse files
committed
Update eager runner and eval llama
1 parent 196499a commit 96ba40b

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

examples/models/llama/eval_llama_lib.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -191,7 +191,7 @@ def gen_eval_wrapper(
191191

192192
pt2e_quant_params, quantizers, quant_dtype = get_quantizer_and_quant_params(args)
193193
# GPTFastEvalWrapper: Create a wrapper around a pre-exported model
194-
manager: LLMEdgeManager = _prepare_for_llama_export(model_name, args)
194+
manager: LLMEdgeManager = _prepare_for_llama_export(args)
195195

196196
if len(quantizers) != 0:
197197
manager = manager.export().pt2e_quantize(quantizers)

examples/models/llama/runner/eager.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ def __init__(self, args):
3838
model_args=model_args,
3939
device="cuda" if torch.cuda.is_available() else "cpu",
4040
)
41-
manager: LLMEdgeManager = _prepare_for_llama_export("llama", args)
41+
manager: LLMEdgeManager = _prepare_for_llama_export(args)
4242
self.model = manager.model.eval().to(device=self.device)
4343

4444
def forward(

0 commit comments

Comments
 (0)