Skip to content

Commit db09613

Browse files
committed
Update on "Move tokenizer.py into extension/llm/tokenizer"
as title Differential Revision: [D59698943](https://our.internmc.facebook.com/intern/diff/D59698943/) [ghstack-poisoned]
2 parents 56eff56 + adf0132 commit db09613

File tree

7 files changed

+15
-9
lines changed

7 files changed

+15
-9
lines changed

.ci/scripts/test_llama.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -171,7 +171,7 @@ $PYTHON_EXECUTABLE -m examples.models.llama2.export_llama ${EXPORT_ARGS}
171171

172172
# Create tokenizer.bin.
173173
echo "Creating tokenizer.bin"
174-
$PYTHON_EXECUTABLE -m examples.models.llama2.tokenizer.tokenizer -t tokenizer.model -o tokenizer.bin
174+
$PYTHON_EXECUTABLE -m extension.llm.tokenizer.tokenizer -t tokenizer.model -o tokenizer.bin
175175

176176

177177
RUNTIME_ARGS="--model_path=${EXPORTED_MODEL_NAME} --tokenizer_path=tokenizer.bin --prompt=Once --temperature=0 --seq_len=10"

backends/vulkan/docs/android_demo.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@ echo '{"dim": 768, "multiple_of": 32, "n_heads": 12, "n_layers": 12, "norm_eps":
6565
Then, create a tokenizer binary file:
6666

6767
```shell
68-
python -m examples.models.llama2.tokenizer.tokenizer -t tokenizer.model -o tokenizer.bin
68+
python -m extension.llm.tokenizer.tokenizer -t tokenizer.model -o tokenizer.bin
6969
```
7070

7171
Finally, export the `stories110M.pt` file into an ExecuTorch program:

examples/models/llama2/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -90,7 +90,7 @@ You can export and run the original Llama 2 7B model.
9090
4. Create tokenizer.bin.
9191
9292
```
93-
python -m examples.models.llama2.tokenizer.tokenizer -t <tokenizer.model> -o tokenizer.bin
93+
python -m extension.llm.tokenizer.tokenizer -t <tokenizer.model> -o tokenizer.bin
9494
```
9595
9696
### Option B: Download and export stories110M model
@@ -113,7 +113,7 @@ If you want to deploy and run a smaller model for educational purposes. From `ex
113113
4. Create tokenizer.bin.
114114
115115
```
116-
python -m examples.models.llama2.tokenizer.tokenizer -t <tokenizer.model> -o tokenizer.bin
116+
python -m extension.llm.tokenizer.tokenizer -t <tokenizer.model> -o tokenizer.bin
117117
```
118118
119119
### Option C: Download and export Llama 3 8B instruct model

examples/models/llama2/eval_llama_lib.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -15,11 +15,11 @@
1515
get_quantizer_and_quant_params,
1616
)
1717
from executorch.examples.models.llama2.tokenizer.tiktoken import Tokenizer as Tiktoken
18-
from executorch.examples.models.llama2.tokenizer.tokenizer import (
19-
Tokenizer as SentencePieceTokenizer,
20-
)
2118

2219
from executorch.extension.llm.export import LLMEdgeManager
20+
from executorch.extension.llm.tokenizer.tokenizer import (
21+
Tokenizer as SentencePieceTokenizer,
22+
)
2323

2424
from lm_eval.api.model import LM
2525

examples/models/llama2/evaluate/eager_eval.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@
1010
import lm_eval
1111
import torch
1212
from executorch.examples.models.llama2.tokenizer.tiktoken import Tokenizer as Tiktoken
13-
from executorch.examples.models.llama2.tokenizer.tokenizer import (
13+
from executorch.extension.llm.tokenizer.tokenizer import (
1414
Tokenizer as SentencePieceTokenizer,
1515
)
1616

examples/qualcomm/llama2/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ wget "https://huggingface.co/karpathy/tinyllamas/resolve/main/stories110M.pt"
2323
wget "https://raw.githubusercontent.com/karpathy/llama2.c/master/tokenizer.model"
2424

2525
# tokenizer.bin:
26-
python -m examples.models.llama2.tokenizer.tokenizer -t tokenizer.model -o tokenizer.bin
26+
python -m extension.llm.tokenizer.tokenizer -t tokenizer.model -o tokenizer.bin
2727

2828
# params.json:
2929
echo '{"dim": 768, "multiple_of": 32, "n_heads": 12, "n_layers": 12, "norm_eps": 1e-05, "vocab_size": 32000}' > params.json

extension/llm/tokenizer/targets.bzl

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,11 @@
11
load("@fbsource//xplat/executorch/build:runtime_wrapper.bzl", "runtime")
22

33
def define_common_targets():
4+
"""Defines targets that should be shared between fbcode and xplat.
5+
6+
The directory containing this targets.bzl file should also contain both
7+
TARGETS and BUCK files that call this function.
8+
"""
49
runtime.python_library(
510
name = "tokenizer_py_lib",
611
srcs = [
@@ -15,6 +20,7 @@ def define_common_targets():
1520
"//bento_kernels/...",
1621
],
1722
_is_external_target = True,
23+
# TODO: Define an external_deps entry for sentencepiece instead of pointing to an fbsource path.
1824
deps = [] if runtime.is_oss else ["fbsource//third-party/pypi/sentencepiece:sentencepiece"],
1925
)
2026

0 commit comments

Comments
 (0)