@@ -2080,9 +2080,11 @@ void llama_vocab::impl::load(llama_model_loader & ml, const LLM_KV & kv) {
2080
2080
2081
2081
std::string model_name;
2082
2082
std::string tokenizer_pre;
2083
+ std::string general_arch;
2083
2084
2084
2085
ml.get_key (LLM_KV_GENERAL_NAME, model_name, false );
2085
2086
ml.get_key (LLM_KV_TOKENIZER_PRE, tokenizer_pre, false );
2087
+ ml.get_key (LLM_KV_GENERAL_ARCHITECTURE, general_arch, false );
2086
2088
2087
2089
// model name to lowercase
2088
2090
std::transform (model_name.begin (), model_name.end (), model_name.begin (),
@@ -2091,8 +2093,11 @@ void llama_vocab::impl::load(llama_model_loader & ml, const LLM_KV & kv) {
2091
2093
}
2092
2094
);
2093
2095
2094
- // set attributes by model/tokenizer name
2095
- if (_contains_any (tokenizer_pre, {" jina-v2-de" , " jina-v2-es" , " jina-v2-code" })) {
2096
+ // set attributes by model/tokenizer/architecture name
2097
+ if (false
2098
+ || _contains_any (tokenizer_pre, {" jina-v2-de" , " jina-v2-es" , " jina-v2-code" })
2099
+ || _contains_any (general_arch, {" nomic-bert-moe" })
2100
+ ) {
2096
2101
_set_token_attr (" <mask>" , LLAMA_TOKEN_ATTR_LSTRIP, true );
2097
2102
} else if (_contains_any (model_name, {" phi-3" , " phi3" })) {
2098
2103
for (auto id : cache_special_tokens) {
0 commit comments