Skip to content

Commit 5834d14

Browse files
committed
Fix encode, remove generated python tokenizer
1 parent 70fd1fe commit 5834d14

File tree

3 files changed

+6
-206
lines changed

3 files changed

+6
-206
lines changed

examples/models/llama/runner/generation.py

Lines changed: 4 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -165,8 +165,7 @@ def text_completion(
165165
This method generates text completion for the provided prompt, employing nucleus sampling to introduce controlled randomness.
166166
"""
167167
return self.generate(
168-
# prompt_tokens=self.tokenizer.encode(prompt, bos=True, eos=False),
169-
prompt_tokens=self.tokenizer.encode(prompt).ids,
168+
prompt_tokens=self.tokenizer.encode(prompt, bos=True, eos=False),
170169
max_seq_len=self.max_seq_len,
171170
temperature=temperature,
172171
top_p=top_p,
@@ -200,10 +199,9 @@ def chat_completion(
200199
prompt = input("Me: ")
201200
while prompt and prompt != exit_prompt:
202201
print("LLM: ", end="", flush=True)
203-
# prompt_tokens = self.tokenizer.encode(
204-
# self._format_prompt(prompt), bos=True, eos=False
205-
# )
206-
prompt_tokens = self.tokenizer.encode(self._format_prompt(prompt)).ids
202+
prompt_tokens = self.tokenizer.encode(
203+
self._format_prompt(prompt), bos=True, eos=False
204+
)
207205
generated_tokens = self.generate(
208206
prompt_tokens=pre_stop_token + prompt_tokens,
209207
max_seq_len=max_seq_len,

extension/llm/tokenizer/hf_tokenizer.py

Lines changed: 0 additions & 199 deletions
This file was deleted.

extension/llm/tokenizer/utils.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,6 @@
88
from typing import Optional
99

1010
from executorch.examples.models.llama.tokenizer.tiktoken import Tokenizer as Tiktoken
11-
from executorch.extension.llm.tokenizer.hf_tokenizer import HFTokenizer
1211
from executorch.extension.llm.tokenizer.tokenizer import (
1312
Tokenizer as SentencePieceTokenizer,
1413
)
@@ -25,6 +24,8 @@ def get_tokenizer(tokenizer_path: str, tokenizer_config_path: Optional[str] = No
2524
tokenizer.n_words = tokenizer.get_vocab_size()
2625
# Keep in line with internal tokenizer apis.
2726
tokenizer.decode_token = lambda token: tokenizer.decode([token])
27+
original_encode = tokenizer.encode
28+
tokenizer.encode = lambda prompt, **kwargs: original_encode(prompt).ids
2829

2930
if tokenizer_config_path:
3031
with open(tokenizer_config_path) as f:

0 commit comments

Comments
 (0)