-
Confusion regarding bin file in README exampleI want to merge my finetuned LoRa adapters into a base model - great I can just use llama-export-lora
But! what confuses me is that the example in the README file uses a bin file for the LoRA adapter . Right now I'm using a older version of So if it's mandatory using a bin version of llora-adapters - How do I make a .bin file? Would really appreciate a full example if possible! What I have right now is the following:
and
|
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 5 replies
-
Sorry the guide has a typo error. The lora must be always gguf: ./bin/llama-export-lora \
-m open-llama-3b-v2-q8_0.gguf \
-o open-llama-3b-v2-q8_0-english2tokipona-chat.gguf \
--lora lora-open-llama-3b-v2-q8_0-english2tokipona-chat-LATEST.gguf Multiple LORA adapters can be applied by passing multiple ./bin/llama-export-lora \
-m your_base_model.gguf \
-o your_merged_model.gguf \
--lora-scaled lora_task_A.gguf 0.5 \
--lora-scaled lora_task_B.gguf 0.5 It's fixed in #8669 |
Beta Was this translation helpful? Give feedback.
-
Thanks @ngxson - I unfortunately get an error in the end - maybe you can spot what's wrong So I converted my lora-adapter to gguf [lora-adapters here]
full text conversion-of-lora-to-gguf.txt Then I tried merging lora_adapter.gguf into the base model, but unfortunately get an error.
|
Beta Was this translation helpful? Give feedback.
-
one last thing @ngxson - seems like my merged model only creates "GGGGGGGGG" - hmmm odd.
But! If I only ran llora seperately with the base model, then it works
|
Beta Was this translation helpful? Give feedback.
Sorry the guide has a typo error. The lora must be always gguf:
Multiple LORA adapters can be applied by passing multiple
--lora FNAME
or--lora-scaled FNAME S
command line parameters:It's fixed in #8669