@@ -605,7 +605,7 @@ void save_as_llama_model(struct llama_vocab * vocab, struct my_llama_model * mod
605605
606606struct train_params get_default_train_params () {
607607 struct train_params params;
608- params.fn_vocab_model = " ggml-vic7b-uncensored-q4_0 .bin" ;
608+ params.fn_vocab_model = " models/ ggml-vocab .bin" ;
609609 params.fn_llama2c_output_model = " ak_llama_model.bin" ;
610610 params.fn_train_data = " shakespeare.txt" ;
611611 params.fn_checkpoint_in = " checkpoint.bin" ;
@@ -658,14 +658,15 @@ void print_usage(int /*argc*/, char ** argv, const struct train_params * params)
658658 fprintf (stderr, " \n " );
659659 fprintf (stderr, " options:\n " );
660660 fprintf (stderr, " -h, --help show this help message and exit\n " );
661- fprintf (stderr, " --vocab-model FNAME model path from which to load vocab (default '%s')\n " , params->fn_vocab_model );
662- fprintf (stderr, " --llama2c-model FNAME model path from which to load Karpathy's llama2.c model\n " );
661+ fprintf (stderr, " --copy- vocab-from- model FNAME model path from which to copy vocab (default '%s')\n " , params->fn_vocab_model );
662+ fprintf (stderr, " --llama2c-model FNAME [REQUIRED] model path from which to load Karpathy's llama2.c model\n " );
663663 fprintf (stderr, " --llama2c-output-model FNAME model path to save the converted llama2.c model (default %s')\n " , params->fn_llama2c_output_model );
664664 fprintf (stderr, " \n " );
665665}
666666
667667bool params_parse (int argc, char ** argv, struct train_params * params) {
668668 bool invalid_param = false ;
669+ bool reqd_param_found = false ;
669670 std::string arg;
670671 struct train_params default_params = get_default_train_params ();
671672 const std::string arg_prefix = " --" ;
@@ -676,7 +677,7 @@ bool params_parse(int argc, char ** argv, struct train_params * params) {
676677 std::replace (arg.begin (), arg.end (), ' _' , ' -' );
677678 }
678679
679- if (arg == " --vocab-model" ) {
680+ if (arg == " --copy- vocab-from -model" ) {
680681 if (++i >= argc) {
681682 invalid_param = true ;
682683 break ;
@@ -687,6 +688,7 @@ bool params_parse(int argc, char ** argv, struct train_params * params) {
687688 invalid_param = true ;
688689 break ;
689690 }
691+ reqd_param_found = true ;
690692 params->fn_llama2c_model = argv[i];
691693 } else if (arg == " --llama2c-output-model" ) {
692694 if (++i >= argc) {
@@ -708,6 +710,11 @@ bool params_parse(int argc, char ** argv, struct train_params * params) {
708710 print_usage (argc, argv, &default_params);
709711 exit (1 );
710712 }
713+ if (!reqd_param_found){
714+ fprintf (stderr, " error: please specify a llama2.c .bin file to be converted with argument --llama2c-model\n " );
715+ print_usage (argc, argv, &default_params);
716+ exit (1 );
717+ }
711718
712719 return true ;
713720}
0 commit comments