Skip to content

Commit 3c0b862

Browse files
committed
rename gpt to common
1 parent 672438d commit 3c0b862

File tree

36 files changed

+677
-677
lines changed

36 files changed

+677
-677
lines changed

common/arg.cpp

Lines changed: 221 additions & 221 deletions
Large diffs are not rendered by default.

common/arg.h

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -18,29 +18,29 @@ struct common_arg {
1818
const char * env = nullptr;
1919
std::string help;
2020
bool is_sparam = false; // is current arg a sampling param?
21-
void (*handler_void) (gpt_params & params) = nullptr;
22-
void (*handler_string) (gpt_params & params, const std::string &) = nullptr;
23-
void (*handler_str_str)(gpt_params & params, const std::string &, const std::string &) = nullptr;
24-
void (*handler_int) (gpt_params & params, int) = nullptr;
21+
void (*handler_void) (common_params & params) = nullptr;
22+
void (*handler_string) (common_params & params, const std::string &) = nullptr;
23+
void (*handler_str_str)(common_params & params, const std::string &, const std::string &) = nullptr;
24+
void (*handler_int) (common_params & params, int) = nullptr;
2525

2626
common_arg(
2727
const std::initializer_list<const char *> & args,
2828
const char * value_hint,
2929
const std::string & help,
30-
void (*handler)(gpt_params & params, const std::string &)
30+
void (*handler)(common_params & params, const std::string &)
3131
) : args(args), value_hint(value_hint), help(help), handler_string(handler) {}
3232

3333
common_arg(
3434
const std::initializer_list<const char *> & args,
3535
const char * value_hint,
3636
const std::string & help,
37-
void (*handler)(gpt_params & params, int)
37+
void (*handler)(common_params & params, int)
3838
) : args(args), value_hint(value_hint), help(help), handler_int(handler) {}
3939

4040
common_arg(
4141
const std::initializer_list<const char *> & args,
4242
const std::string & help,
43-
void (*handler)(gpt_params & params)
43+
void (*handler)(common_params & params)
4444
) : args(args), help(help), handler_void(handler) {}
4545

4646
// support 2 values for arg
@@ -49,7 +49,7 @@ struct common_arg {
4949
const char * value_hint,
5050
const char * value_hint_2,
5151
const std::string & help,
52-
void (*handler)(gpt_params & params, const std::string &, const std::string &)
52+
void (*handler)(common_params & params, const std::string &, const std::string &)
5353
) : args(args), value_hint(value_hint), value_hint_2(value_hint_2), help(help), handler_str_str(handler) {}
5454

5555
common_arg & set_examples(std::initializer_list<enum llama_example> examples);
@@ -61,17 +61,17 @@ struct common_arg {
6161
std::string to_string();
6262
};
6363

64-
struct gpt_params_context {
64+
struct common_params_context {
6565
enum llama_example ex = LLAMA_EXAMPLE_COMMON;
66-
gpt_params & params;
66+
common_params & params;
6767
std::vector<common_arg> options;
6868
void(*print_usage)(int, char **) = nullptr;
69-
gpt_params_context(gpt_params & params) : params(params) {}
69+
common_params_context(common_params & params) : params(params) {}
7070
};
7171

7272
// parse input arguments from CLI
7373
// if one argument has invalid value, it will automatically display usage of the specific argument (and not the full usage message)
74-
bool gpt_params_parse(int argc, char ** argv, gpt_params & params, llama_example ex, void(*print_usage)(int, char **) = nullptr);
74+
bool common_params_parse(int argc, char ** argv, common_params & params, llama_example ex, void(*print_usage)(int, char **) = nullptr);
7575

7676
// function to be used by test-arg-parser
77-
gpt_params_context gpt_params_parser_init(gpt_params & params, llama_example ex, void(*print_usage)(int, char **) = nullptr);
77+
common_params_context common_params_parser_init(common_params & params, llama_example ex, void(*print_usage)(int, char **) = nullptr);

common/common.cpp

Lines changed: 10 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -362,10 +362,10 @@ bool parse_cpu_mask(const std::string & mask, bool (&boolmask)[GGML_MAX_N_THREAD
362362
return true;
363363
}
364364

365-
void gpt_init() {
365+
void common_init() {
366366
llama_log_set([](ggml_log_level level, const char * text, void * /*user_data*/) {
367-
if (LOG_DEFAULT_LLAMA <= gpt_log_verbosity_thold) {
368-
gpt_log_add(gpt_log_main(), level, "%s", text);
367+
if (LOG_DEFAULT_LLAMA <= common_log_verbosity_thold) {
368+
common_log_add(common_log_main(), level, "%s", text);
369369
}
370370
}, NULL);
371371

@@ -378,7 +378,7 @@ void gpt_init() {
378378
LOG_INF("build: %d (%s) with %s for %s%s\n", LLAMA_BUILD_NUMBER, LLAMA_COMMIT, LLAMA_COMPILER, LLAMA_BUILD_TARGET, build_type);
379379
}
380380

381-
std::string gpt_params_get_system_info(const gpt_params & params) {
381+
std::string common_params_get_system_info(const common_params & params) {
382382
std::ostringstream os;
383383

384384
os << "system_info: n_threads = " << params.cpuparams.n_threads;
@@ -819,9 +819,9 @@ std::string fs_get_cache_file(const std::string & filename) {
819819
//
820820
// Model utils
821821
//
822-
struct common_init_result llama_init_from_gpt_params(gpt_params & params) {
822+
struct common_init_result common_init_from_common_params(common_params & params) {
823823
common_init_result iparams;
824-
auto mparams = common_model_params_from_gpt_params(params);
824+
auto mparams = common_model_params_from_common_params(params);
825825

826826
llama_model * model = nullptr;
827827

@@ -863,7 +863,7 @@ struct common_init_result llama_init_from_gpt_params(gpt_params & params) {
863863
}
864864
}
865865

866-
auto cparams = common_context_params_from_gpt_params(params);
866+
auto cparams = common_context_params_from_common_params(params);
867867

868868
llama_context * lctx = llama_new_context_with_model(model, cparams);
869869
if (lctx == NULL) {
@@ -970,7 +970,7 @@ void common_lora_adapters_apply(struct llama_context * ctx, std::vector<common_l
970970
}
971971
}
972972

973-
struct llama_model_params common_model_params_from_gpt_params(const gpt_params & params) {
973+
struct llama_model_params common_model_params_from_common_params(const common_params & params) {
974974
auto mparams = llama_model_default_params();
975975

976976
if (params.n_gpu_layers != -1) {
@@ -1022,7 +1022,7 @@ static ggml_type kv_cache_type_from_str(const std::string & s) {
10221022
throw std::runtime_error("Invalid cache type: " + s);
10231023
}
10241024

1025-
struct llama_context_params common_context_params_from_gpt_params(const gpt_params & params) {
1025+
struct llama_context_params common_context_params_from_common_params(const common_params & params) {
10261026
auto cparams = llama_context_default_params();
10271027

10281028
cparams.n_ctx = params.n_ctx;
@@ -1946,7 +1946,7 @@ void yaml_dump_string_multiline(FILE * stream, const char * prop_name, const cha
19461946
}
19471947
}
19481948

1949-
void yaml_dump_non_result_info(FILE * stream, const gpt_params & params, const llama_context * lctx,
1949+
void yaml_dump_non_result_info(FILE * stream, const common_params & params, const llama_context * lctx,
19501950
const std::string & timestamp, const std::vector<int> & prompt_tokens, const char * model_desc) {
19511951
const auto & sparams = params.sparams;
19521952

common/common.h

Lines changed: 24 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -82,14 +82,14 @@ enum llama_example {
8282
LLAMA_EXAMPLE_COUNT,
8383
};
8484

85-
enum gpt_sampler_type {
86-
GPT_SAMPLER_TYPE_NONE = 0,
87-
GPT_SAMPLER_TYPE_TOP_K = 1,
88-
GPT_SAMPLER_TYPE_TOP_P = 2,
89-
GPT_SAMPLER_TYPE_MIN_P = 3,
90-
GPT_SAMPLER_TYPE_TFS_Z = 4,
91-
GPT_SAMPLER_TYPE_TYPICAL_P = 5,
92-
GPT_SAMPLER_TYPE_TEMPERATURE = 6,
85+
enum common_sampler_type {
86+
COMMON_SAMPLER_TYPE_NONE = 0,
87+
COMMON_SAMPLER_TYPE_TOP_K = 1,
88+
COMMON_SAMPLER_TYPE_TOP_P = 2,
89+
COMMON_SAMPLER_TYPE_MIN_P = 3,
90+
COMMON_SAMPLER_TYPE_TFS_Z = 4,
91+
COMMON_SAMPLER_TYPE_TYPICAL_P = 5,
92+
COMMON_SAMPLER_TYPE_TEMPERATURE = 6,
9393
};
9494

9595
// dimensionality reduction methods, used by cvector-generator
@@ -99,7 +99,7 @@ enum dimre_method {
9999
};
100100

101101
// sampler parameters
102-
struct gpt_sampler_params {
102+
struct common_sampler_params {
103103
uint32_t seed = LLAMA_DEFAULT_SEED; // the seed used to initialize llama_sampler
104104

105105
int32_t n_prev = 64; // number of previous tokens to remember
@@ -124,13 +124,13 @@ struct gpt_sampler_params {
124124
bool ignore_eos = false;
125125
bool no_perf = false; // disable performance metrics
126126

127-
std::vector<enum gpt_sampler_type> samplers = {
128-
GPT_SAMPLER_TYPE_TOP_K,
129-
GPT_SAMPLER_TYPE_TFS_Z,
130-
GPT_SAMPLER_TYPE_TYPICAL_P,
131-
GPT_SAMPLER_TYPE_TOP_P,
132-
GPT_SAMPLER_TYPE_MIN_P,
133-
GPT_SAMPLER_TYPE_TEMPERATURE
127+
std::vector<enum common_sampler_type> samplers = {
128+
COMMON_SAMPLER_TYPE_TOP_K,
129+
COMMON_SAMPLER_TYPE_TFS_Z,
130+
COMMON_SAMPLER_TYPE_TYPICAL_P,
131+
COMMON_SAMPLER_TYPE_TOP_P,
132+
COMMON_SAMPLER_TYPE_MIN_P,
133+
COMMON_SAMPLER_TYPE_TEMPERATURE
134134
};
135135

136136
std::string grammar; // optional BNF-like grammar to constrain sampling
@@ -141,7 +141,7 @@ struct gpt_sampler_params {
141141
std::string print() const;
142142
};
143143

144-
struct gpt_params {
144+
struct common_params {
145145
int32_t n_predict = -1; // new tokens to predict
146146
int32_t n_ctx = 0; // context size
147147
int32_t n_batch = 2048; // logical batch size for prompt processing (must be >=32 to use BLAS)
@@ -183,7 +183,7 @@ struct gpt_params {
183183
enum llama_pooling_type pooling_type = LLAMA_POOLING_TYPE_UNSPECIFIED; // pooling type for embeddings
184184
enum llama_attention_type attention_type = LLAMA_ATTENTION_TYPE_UNSPECIFIED; // attention type for embeddings
185185

186-
struct gpt_sampler_params sparams;
186+
struct common_sampler_params sparams;
187187

188188
std::string model = ""; // model path // NOLINT
189189
std::string model_draft = ""; // draft model for speculative decoding // NOLINT
@@ -348,9 +348,9 @@ struct gpt_params {
348348

349349
// call once at the start of a program if it uses libcommon
350350
// initializes the logging system and prints info about the build
351-
void gpt_init();
351+
void common_init();
352352

353-
std::string gpt_params_get_system_info(const gpt_params & params);
353+
std::string common_params_get_system_info(const common_params & params);
354354

355355
bool parse_cpu_range(const std::string& range, bool(&boolmask)[GGML_MAX_N_THREADS]);
356356
bool parse_cpu_mask(const std::string& mask, bool(&boolmask)[GGML_MAX_N_THREADS]);
@@ -410,10 +410,10 @@ struct common_init_result {
410410
std::vector<common_lora_adapter_container> lora_adapters;
411411
};
412412

413-
struct common_init_result llama_init_from_gpt_params(gpt_params & params);
413+
struct common_init_result common_init_from_common_params(common_params & params);
414414

415-
struct llama_model_params common_model_params_from_gpt_params (const gpt_params & params);
416-
struct llama_context_params common_context_params_from_gpt_params (const gpt_params & params);
415+
struct llama_model_params common_model_params_from_common_params (const common_params & params);
416+
struct llama_context_params common_context_params_from_common_params(const common_params & params);
417417
struct ggml_threadpool_params ggml_threadpool_params_from_cpu_params(const cpu_params & params);
418418

419419
struct llama_model * common_load_model_from_url(const char * model_url, const char * path_model, const char * hf_token, const struct llama_model_params & params);
@@ -554,5 +554,5 @@ void yaml_dump_vector_int (FILE * stream, const char * prop_name, const std
554554
void yaml_dump_string_multiline(FILE * stream, const char * prop_name, const char * data);
555555

556556
void yaml_dump_non_result_info(
557-
FILE * stream, const gpt_params & params, const llama_context * lctx,
557+
FILE * stream, const common_params & params, const llama_context * lctx,
558558
const std::string & timestamp, const std::vector<int> & prompt_tokens, const char * model_desc);

0 commit comments

Comments
 (0)