Skip to content

Commit fea95c6

Browse files
authored
fix convert.py for codellama, add llama 34B to the list of recognized models (#2768)
1 parent ef955fb commit fea95c6

File tree

2 files changed

+4
-1
lines changed

2 files changed

+4
-1
lines changed

convert.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -191,7 +191,7 @@ def loadHFTransformerJson(model: 'LazyModel', config_path: 'Path') -> 'Params':
191191
def loadOriginalParamsJson(model: 'LazyModel', config_path: 'Path') -> 'Params':
192192
config = json.load(open(config_path))
193193

194-
n_vocab = config["vocab_size"]
194+
n_vocab = config["vocab_size"] if "vocab_size" in config else -1
195195
n_embd = config["dim"]
196196
n_layer = config["n_layers"]
197197
n_mult = config["multiple_of"]

llama.cpp

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -827,6 +827,7 @@ enum e_model {
827827
MODEL_7B,
828828
MODEL_13B,
829829
MODEL_30B,
830+
MODEL_34B,
830831
MODEL_40B,
831832
MODEL_65B,
832833
MODEL_70B,
@@ -1518,6 +1519,7 @@ static const char * llama_model_type_name(e_model type) {
15181519
case MODEL_7B: return "7B";
15191520
case MODEL_13B: return "13B";
15201521
case MODEL_30B: return "30B";
1522+
case MODEL_34B: return "34B";
15211523
case MODEL_40B: return "40B";
15221524
case MODEL_65B: return "65B";
15231525
case MODEL_70B: return "70B";
@@ -1590,6 +1592,7 @@ static void llm_load_hparams(
15901592
case 26: model.type = e_model::MODEL_3B; break;
15911593
case 32: model.type = e_model::MODEL_7B; break;
15921594
case 40: model.type = e_model::MODEL_13B; break;
1595+
case 48: model.type = e_model::MODEL_34B; break;
15931596
case 60: model.type = e_model::MODEL_30B; break;
15941597
case 80: model.type = hparams.n_head == hparams.n_head_kv ? e_model::MODEL_65B : e_model::MODEL_70B; break;
15951598
default: model.type = e_model::MODEL_UNKNOWN;

0 commit comments

Comments
 (0)