@@ -605,7 +605,7 @@ void save_as_llama_model(struct llama_vocab * vocab, struct my_llama_model * mod
605
605
606
606
struct train_params get_default_train_params () {
607
607
struct train_params params;
608
- params.fn_vocab_model = " ggml-vic7b-uncensored-q4_0 .bin" ;
608
+ params.fn_vocab_model = " models/ ggml-vocab .bin" ;
609
609
params.fn_llama2c_output_model = " ak_llama_model.bin" ;
610
610
params.fn_train_data = " shakespeare.txt" ;
611
611
params.fn_checkpoint_in = " checkpoint.bin" ;
@@ -658,14 +658,15 @@ void print_usage(int /*argc*/, char ** argv, const struct train_params * params)
658
658
fprintf (stderr, " \n " );
659
659
fprintf (stderr, " options:\n " );
660
660
fprintf (stderr, " -h, --help show this help message and exit\n " );
661
- fprintf (stderr, " --vocab-model FNAME model path from which to load vocab (default '%s')\n " , params->fn_vocab_model );
662
- fprintf (stderr, " --llama2c-model FNAME model path from which to load Karpathy's llama2.c model\n " );
661
+ fprintf (stderr, " --copy- vocab-from- model FNAME model path from which to copy vocab (default '%s')\n " , params->fn_vocab_model );
662
+ fprintf (stderr, " --llama2c-model FNAME [REQUIRED] model path from which to load Karpathy's llama2.c model\n " );
663
663
fprintf (stderr, " --llama2c-output-model FNAME model path to save the converted llama2.c model (default %s')\n " , params->fn_llama2c_output_model );
664
664
fprintf (stderr, " \n " );
665
665
}
666
666
667
667
bool params_parse (int argc, char ** argv, struct train_params * params) {
668
668
bool invalid_param = false ;
669
+ bool reqd_param_found = false ;
669
670
std::string arg;
670
671
struct train_params default_params = get_default_train_params ();
671
672
const std::string arg_prefix = " --" ;
@@ -676,7 +677,7 @@ bool params_parse(int argc, char ** argv, struct train_params * params) {
676
677
std::replace (arg.begin (), arg.end (), ' _' , ' -' );
677
678
}
678
679
679
- if (arg == " --vocab-model" ) {
680
+ if (arg == " --copy- vocab-from -model" ) {
680
681
if (++i >= argc) {
681
682
invalid_param = true ;
682
683
break ;
@@ -687,6 +688,7 @@ bool params_parse(int argc, char ** argv, struct train_params * params) {
687
688
invalid_param = true ;
688
689
break ;
689
690
}
691
+ reqd_param_found = true ;
690
692
params->fn_llama2c_model = argv[i];
691
693
} else if (arg == " --llama2c-output-model" ) {
692
694
if (++i >= argc) {
@@ -708,6 +710,11 @@ bool params_parse(int argc, char ** argv, struct train_params * params) {
708
710
print_usage (argc, argv, &default_params);
709
711
exit (1 );
710
712
}
713
+ if (!reqd_param_found){
714
+ fprintf (stderr, " error: please specify a llama2.c .bin file to be converted with argument --llama2c-model\n " );
715
+ print_usage (argc, argv, &default_params);
716
+ exit (1 );
717
+ }
711
718
712
719
return true ;
713
720
}
0 commit comments