@@ -576,7 +576,7 @@ struct llama_model_loader {
576
576
struct ggml_context * ggml_ctx = NULL ;
577
577
std::unique_ptr<llama_mmap> mapping;
578
578
579
- llama_model_loader (const std::string & fname_base, bool use_mmap, bool vocab_only ) {
579
+ llama_model_loader (const std::string & fname_base, bool use_mmap) {
580
580
file_loader = std::unique_ptr<llama_file_loader>(new llama_file_loader (fname_base.c_str (), tensors_map));
581
581
if (!llama_mmap::SUPPORTED) {
582
582
use_mmap = false ;
@@ -921,7 +921,7 @@ static void llama_model_load_internal(
921
921
922
922
model.t_start_us = ggml_time_us ();
923
923
924
- std::unique_ptr<llama_model_loader> ml (new llama_model_loader (fname, use_mmap, vocab_only ));
924
+ std::unique_ptr<llama_model_loader> ml (new llama_model_loader (fname, use_mmap));
925
925
926
926
vocab = std::move (ml->file_loader ->vocab );
927
927
model.hparams = ml->file_loader ->hparams ;
@@ -2304,8 +2304,7 @@ static void llama_model_quantize_internal(const std::string & fname_inp, const s
2304
2304
nthread = std::thread::hardware_concurrency ();
2305
2305
}
2306
2306
2307
- std::unique_ptr<llama_model_loader> model_loader (new llama_model_loader (fname_inp, /* use_mmap*/ false ,
2308
- /* vocab_only*/ false ));
2307
+ std::unique_ptr<llama_model_loader> model_loader (new llama_model_loader (fname_inp, /* use_mmap*/ false ));
2309
2308
llama_file_saver file_saver (fname_out.c_str (), model_loader->file_loader .get (), params->ftype );
2310
2309
2311
2310
#ifdef GGML_USE_K_QUANTS
@@ -2738,7 +2737,7 @@ int llama_apply_lora_from_file_internal(const struct llama_model & model, const
2738
2737
llama_buffer base_buf;
2739
2738
if (path_base_model) {
2740
2739
fprintf (stderr, " %s: loading base model from '%s'\n " , __func__, path_base_model);
2741
- model_loader.reset (new llama_model_loader (path_base_model, /* use_mmap*/ true , /* vocab_only */ false ));
2740
+ model_loader.reset (new llama_model_loader (path_base_model, /* use_mmap*/ true ));
2742
2741
2743
2742
size_t ctx_size;
2744
2743
size_t mmapped_size;
0 commit comments