Skip to content

Commit e4d0d97

Browse files
committed
Fix the missing vocab params in llama.py
1 parent 52327de commit e4d0d97

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

llama_cpp/llama.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -478,10 +478,10 @@ def free_lora_adapter():
478478
bos_token_id = self.token_bos()
479479

480480
eos_token = (
481-
self._model.token_get_text(eos_token_id) if eos_token_id != -1 else ""
481+
self._model.token_get_text(self._vocab, eos_token_id) if eos_token_id != -1 else ""
482482
)
483483
bos_token = (
484-
self._model.token_get_text(bos_token_id) if bos_token_id != -1 else ""
484+
self._model.token_get_text(self._vocab, bos_token_id) if bos_token_id != -1 else ""
485485
)
486486

487487
# Unfortunately the llama.cpp API does not return metadata arrays, so we can't get template names from tokenizer.chat_templates
@@ -1237,7 +1237,7 @@ def _create_completion(
12371237

12381238
if prompt_tokens[:2] == [self.token_bos()] * 2:
12391239
warnings.warn(
1240-
f'Detected duplicate leading "{self._model.token_get_text(self.token_bos())}" in prompt, this will likely reduce response quality, consider removing it...',
1240+
f'Detected duplicate leading "{self._model.token_get_text(self._vocab, self.token_bos())}" in prompt, this will likely reduce response quality, consider removing it...',
12411241
RuntimeWarning,
12421242
)
12431243

0 commit comments

Comments
 (0)