We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 20b8f83 commit 7315894Copy full SHA for 7315894
examples/models/llama2/eval_llama_lib.py
@@ -44,8 +44,12 @@ def __init__(
44
super().__init__(None, tokenizer, max_seq_length)
45
self._model = model # Expects model to be path to a .pte file
46
47
+ from executorch.extension.pybindings import portable_lib # noqa
48
from executorch.extension.pybindings.portable_lib import _load_for_executorch
-
49
+ # Note: import this after portable_lib
50
+ from executorch.extension.llm.custom_ops import sdpa_with_kv_cache # noqa
51
+ from executorch.kernels import quantized
52
+
53
self._et_model = _load_for_executorch(self._model)
54
self._use_kv_cache = self._et_model.run_method("use_kv_cache")[0]
55
0 commit comments