You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
> Vote for which quantization type provides better responses, all other parameters being the same.
15
-
16
11
### Recent API changes
17
12
18
13
-[2024 Mar 8]`llama_kv_cache_seq_rm()` returns a `bool` instead of `void`, and new `llama_n_max_seq()` returns the upper limit of acceptable `seq_id` in batches (relevant when dealing with multiple sequences) https://github.com/ggerganov/llama.cpp/pull/5328
@@ -21,6 +16,8 @@ Inference of Meta's [LLaMA](https://arxiv.org/abs/2302.13971) model (and others)
21
16
22
17
### Hot topics
23
18
19
+
- Looking for contributions to add Deepseek support: https://github.com/ggerganov/llama.cpp/issues/5981
0 commit comments