Skip to content

Releases: ggml-org/llama.cpp

b4280

07 Dec 10:05
3df784b
Compare
Choose a tag to compare
Vulkan: VK_KHR_cooperative_matrix support to speed up prompt processi…

b4279

07 Dec 08:33
86a1934
Compare
Choose a tag to compare
metal : Extend how Llama.cpp locates metal resources (#10676)

* metal : Extend how Llama.cpp locates metal resources (#10675)

  * It searches the resource file in the directory where the current
    binary is located as well.
  * Resolves symbolic links.

Rationale:

When we plug this dependency into a Bazel build and run it in the
context of Bazel (e.g. testing):

  * the execution directory is often very different from where the files
    are located and no direct control over this (Bazel sandboxing),
  * the Bazel sandbox often use symbolic links to make files available.

With this patch, we can have the resource file added to the target,
can build and run tests in the context of Bazel.

* Update ggml/src/ggml-metal/ggml-metal.m

Co-authored-by: Georgi Gerganov <[email protected]>

* Update ggml/src/ggml-metal/ggml-metal.m

Co-authored-by: Georgi Gerganov <[email protected]>

---------

Co-authored-by: Georgi Gerganov <[email protected]>

b4276

06 Dec 13:10
f162d45
Compare
Choose a tag to compare
common : bring back --no-warmup to server (#10686)

b4273

05 Dec 19:59
c9c6e01
Compare
Choose a tag to compare
vulkan: Add VK_NV_cooperative_matrix2 support for mul_mat and flash a…

b4272

05 Dec 19:25
6fe6247
Compare
Choose a tag to compare
llama : add Minerva 7B model support (#10673)

* Support for Minerva 7B

* Update convert_hf_to_gguf_update.py

b4271

05 Dec 12:12
0cd182e
Compare
Choose a tag to compare
sync : ggml

b4267

04 Dec 23:16
f112d19
Compare
Choose a tag to compare
Update deprecation-warning.cpp (#10619)

Fixed Path Separator Handling for Cross-Platform Support (Windows File Systems)

b4266

04 Dec 21:22
1da7b76
Compare
Choose a tag to compare
server : fix speculative decoding with context shift (#10641)

* server : fix speculative decoding with context shift

ggml-ci

* server : take into account speculative limits

ggml-ci

* server : add tests

b4265

04 Dec 14:49
59f4db1
Compare
Choose a tag to compare
ggml : add predefined list of CPU backend variants to build (#10626)

* ggml : add predefined list of CPU backend variants to build

* update CPU dockerfiles

b4262

04 Dec 10:25
8d0cfd5
Compare
Choose a tag to compare
llama: Support MiniCPM-1B (with & w/o longrope) (#10559)