-
Notifications
You must be signed in to change notification settings - Fork 12.2k
vocab : BailingMoE : change possessive quantifiers to greedy #12677
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Did you do tokenizer tests to make sure the results match with the reference tokenizer? |
I did some basic tests, but I will admit that I'm not entirely sure how to set up a proper test, any pointers? |
You need to run: python convert_hf_to_gguf_update.py <hf_token> This will download reference tokenizers for all models to After that, create a "vocab-only" GGUF model: # this is for llama - update to create one for the Bailing model
python3 convert_hf_to_gguf.py models/tokenizers/llama-spm/ --outfile models/ggml-vocab-llama-spm.gguf --vocab-only Run the |
Ah, I missed the relationship between these files, I see now, added for future tests.
|
@ggerganov gentle ping :) |
Nice, but I didn't mean to commit the generated test files. At some point we stopped source controlling them because they add non negligible data (in this case 5MB). The idea is to just generate and test them locally. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Merge after removing the vocab files.
False alarms all around, was accidentally using an older build on my GPU... ignore me :) thanks so much for the fix! |
The possessive quantifiers are causing weird issues, and atomic grouping does not seem to be supported, so revert to greedy.
See following reports:
@bartowski1182 @nicoboss