Releases: ggml-org/llama.cpp
Releases · ggml-org/llama.cpp
b4280
Vulkan: VK_KHR_cooperative_matrix support to speed up prompt processi…
b4279
metal : Extend how Llama.cpp locates metal resources (#10676) * metal : Extend how Llama.cpp locates metal resources (#10675) * It searches the resource file in the directory where the current binary is located as well. * Resolves symbolic links. Rationale: When we plug this dependency into a Bazel build and run it in the context of Bazel (e.g. testing): * the execution directory is often very different from where the files are located and no direct control over this (Bazel sandboxing), * the Bazel sandbox often use symbolic links to make files available. With this patch, we can have the resource file added to the target, can build and run tests in the context of Bazel. * Update ggml/src/ggml-metal/ggml-metal.m Co-authored-by: Georgi Gerganov <[email protected]> * Update ggml/src/ggml-metal/ggml-metal.m Co-authored-by: Georgi Gerganov <[email protected]> --------- Co-authored-by: Georgi Gerganov <[email protected]>
b4276
common : bring back --no-warmup to server (#10686)
b4273
vulkan: Add VK_NV_cooperative_matrix2 support for mul_mat and flash a…
b4272
llama : add Minerva 7B model support (#10673) * Support for Minerva 7B * Update convert_hf_to_gguf_update.py
b4271
sync : ggml
b4267
Update deprecation-warning.cpp (#10619) Fixed Path Separator Handling for Cross-Platform Support (Windows File Systems)
b4266
server : fix speculative decoding with context shift (#10641) * server : fix speculative decoding with context shift ggml-ci * server : take into account speculative limits ggml-ci * server : add tests
b4265
ggml : add predefined list of CPU backend variants to build (#10626) * ggml : add predefined list of CPU backend variants to build * update CPU dockerfiles
b4262
llama: Support MiniCPM-1B (with & w/o longrope) (#10559)