File tree Expand file tree Collapse file tree 1 file changed +9
-3
lines changed Expand file tree Collapse file tree 1 file changed +9
-3
lines changed Original file line number Diff line number Diff line change 5
5
6
6
Inference of [ LLaMA] ( https://arxiv.org/abs/2302.13971 ) model in pure C/C++
7
7
8
+ ---
9
+
10
+ ** TEMPORARY NOTICE:**
11
+ Big code change incoming: https://github.com/ggerganov/llama.cpp/pull/370
12
+
13
+ Do not merge stuff until we merge this. Probably merge will happen on March 22 ~ 6:00am UTC
14
+
15
+ ---
16
+
8
17
** Hot topics:**
9
18
10
19
- [ Added Alpaca support] ( https://github.com/ggerganov/llama.cpp#instruction-mode-with-alpaca )
11
20
- Cache input prompts for faster initialization: https://github.com/ggerganov/llama.cpp/issues/64
12
21
- Create a ` llama.cpp ` logo: https://github.com/ggerganov/llama.cpp/issues/105
13
22
14
- ** TEMPORARY NOTICE:**
15
- If you're updating to the latest master, you will need to regenerate your model files as the format has changed.
16
-
17
23
## Description
18
24
19
25
The main goal is to run the model using 4-bit quantization on a MacBook
You can’t perform that action at this time.
0 commit comments