Skip to content

Commit 220d931

Browse files
ftgreatldwang
andauthored
readme : add Aquila-7B model series to supported models (#2487)
* support bpe tokenizer in convert Signed-off-by: ldwang <[email protected]> * support bpe tokenizer in convert Signed-off-by: ldwang <[email protected]> * support bpe tokenizer in convert, fix Signed-off-by: ldwang <[email protected]> * Add Aquila-7B models in README.md Signed-off-by: ldwang <[email protected]> * Up Aquila-7B models in README.md Signed-off-by: ldwang <[email protected]> --------- Signed-off-by: ldwang <[email protected]> Co-authored-by: ldwang <[email protected]>
1 parent 81844fb commit 220d931

File tree

1 file changed

+7
-0
lines changed

1 file changed

+7
-0
lines changed

README.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -88,6 +88,7 @@ as the main playground for developing new features for the [ggml](https://github
8888
- [X] [Pygmalion 7B / Metharme 7B](#using-pygmalion-7b--metharme-7b)
8989
- [X] [WizardLM](https://github.com/nlpxucan/WizardLM)
9090
- [X] [Baichuan-7B](https://huggingface.co/baichuan-inc/baichuan-7B) and its derivations (such as [baichuan-7b-sft](https://huggingface.co/hiyouga/baichuan-7b-sft))
91+
- [X] [Aquila-7B](https://huggingface.co/BAAI/Aquila-7B) / [AquilaChat-7B](https://huggingface.co/BAAI/AquilaChat-7B)
9192

9293
**Bindings:**
9394

@@ -492,13 +493,19 @@ Building the program with BLAS support may lead to some performance improvements
492493
# obtain the original LLaMA model weights and place them in ./models
493494
ls ./models
494495
65B 30B 13B 7B tokenizer_checklist.chk tokenizer.model
496+
# [Optional] for models using BPE tokenizers
497+
ls ./models
498+
65B 30B 13B 7B vocab.json
495499
496500
# install Python dependencies
497501
python3 -m pip install -r requirements.txt
498502
499503
# convert the 7B model to ggml FP16 format
500504
python3 convert.py models/7B/
501505
506+
# [Optional] for models using BPE tokenizers
507+
python convert.py models/7B/ --vocabtype bpe
508+
502509
# quantize the model to 4-bits (using q4_0 method)
503510
./quantize ./models/7B/ggml-model-f16.bin ./models/7B/ggml-model-q4_0.bin q4_0
504511

0 commit comments

Comments
 (0)