@@ -39,6 +39,10 @@ bool gpt_params_parse(int argc, char ** argv, gpt_params & params) {
39
39
params.top_p = std::stof (argv[++i]);
40
40
} else if (arg == " --temp" ) {
41
41
params.temp = std::stof (argv[++i]);
42
+ } else if (arg == " --repeat-last-n" ) {
43
+ params.repeat_last_n = std::stof (argv[++i]);
44
+ } else if (arg == " --repeat-penalty" ) {
45
+ params.repeat_penalty = std::stof (argv[++i]);
42
46
} else if (arg == " -b" || arg == " --batch_size" ) {
43
47
params.n_batch = std::stoi (argv[++i]);
44
48
} else if (arg == " -m" || arg == " --model" ) {
@@ -90,6 +94,8 @@ void gpt_print_usage(int /*argc*/, char ** argv, const gpt_params & params) {
90
94
fprintf (stderr, " --top_k N top-k sampling (default: %d)\n " , params.top_k );
91
95
fprintf (stderr, " --top_p N top-p sampling (default: %.1f)\n " , params.top_p );
92
96
fprintf (stderr, " --temp N temperature (default: %.1f)\n " , params.temp );
97
+ fprintf (stderr, " --repeat-last-n N last n tokens to consider for penalize (default: %d, 0 = disabled)\n " , params.repeat_last_n );
98
+ fprintf (stderr, " --repeat-penalty N penalize repeat sequence of tokens (default: %.2f, 1.0 = disabled)\n " , (double )params.repeat_penalty );
93
99
fprintf (stderr, " -b N, --batch_size N batch size for prompt processing (default: %d)\n " , params.n_batch );
94
100
fprintf (stderr, " -m FNAME, --model FNAME\n " );
95
101
fprintf (stderr, " model path (default: %s)\n " , params.model .c_str ());
0 commit comments