Skip to content

Commit aab15de

Browse files
author
Aniket
committed
commandline argument changes for clarity.
Changed the parameter to load the vocab model to --copy-vocab-from-model Made the default vocab model to load from models/ggml-vocal.bin Made sure a llama2.c model is provided, or else exit with usage. Updated README.md for better use.
1 parent db5d7ab commit aab15de

File tree

1 file changed

+11
-4
lines changed

1 file changed

+11
-4
lines changed

examples/convert-llama2c-to-ggml/convert-llama2c-to-ggml.cpp

Lines changed: 11 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -605,7 +605,7 @@ void save_as_llama_model(struct llama_vocab * vocab, struct my_llama_model * mod
605605

606606
struct train_params get_default_train_params() {
607607
struct train_params params;
608-
params.fn_vocab_model = "ggml-vic7b-uncensored-q4_0.bin";
608+
params.fn_vocab_model = "models/ggml-vocab.bin";
609609
params.fn_llama2c_output_model = "ak_llama_model.bin";
610610
params.fn_train_data = "shakespeare.txt";
611611
params.fn_checkpoint_in = "checkpoint.bin";
@@ -658,14 +658,15 @@ void print_usage(int /*argc*/, char ** argv, const struct train_params * params)
658658
fprintf(stderr, "\n");
659659
fprintf(stderr, "options:\n");
660660
fprintf(stderr, " -h, --help show this help message and exit\n");
661-
fprintf(stderr, " --vocab-model FNAME model path from which to load vocab (default '%s')\n", params->fn_vocab_model);
662-
fprintf(stderr, " --llama2c-model FNAME model path from which to load Karpathy's llama2.c model\n");
661+
fprintf(stderr, " --copy-vocab-from-model FNAME model path from which to copy vocab (default '%s')\n", params->fn_vocab_model);
662+
fprintf(stderr, " --llama2c-model FNAME [REQUIRED] model path from which to load Karpathy's llama2.c model\n");
663663
fprintf(stderr, " --llama2c-output-model FNAME model path to save the converted llama2.c model (default %s')\n", params->fn_llama2c_output_model);
664664
fprintf(stderr, "\n");
665665
}
666666

667667
bool params_parse(int argc, char ** argv, struct train_params * params) {
668668
bool invalid_param = false;
669+
bool reqd_param_found = false;
669670
std::string arg;
670671
struct train_params default_params = get_default_train_params();
671672
const std::string arg_prefix = "--";
@@ -676,7 +677,7 @@ bool params_parse(int argc, char ** argv, struct train_params * params) {
676677
std::replace(arg.begin(), arg.end(), '_', '-');
677678
}
678679

679-
if (arg == "--vocab-model") {
680+
if (arg == "--copy-vocab-from-model") {
680681
if (++i >= argc) {
681682
invalid_param = true;
682683
break;
@@ -687,6 +688,7 @@ bool params_parse(int argc, char ** argv, struct train_params * params) {
687688
invalid_param = true;
688689
break;
689690
}
691+
reqd_param_found = true;
690692
params->fn_llama2c_model = argv[i];
691693
} else if (arg == "--llama2c-output-model") {
692694
if (++i >= argc) {
@@ -708,6 +710,11 @@ bool params_parse(int argc, char ** argv, struct train_params * params) {
708710
print_usage(argc, argv, &default_params);
709711
exit(1);
710712
}
713+
if (!reqd_param_found){
714+
fprintf(stderr, "error: please specify a llama2.c .bin file to be converted with argument --llama2c-model\n");
715+
print_usage(argc, argv, &default_params);
716+
exit(1);
717+
}
711718

712719
return true;
713720
}

0 commit comments

Comments
 (0)