Skip to content

Commit d97b1ff

Browse files
authored
Merge pull request #475 from janhq/473-chore-pump-llama.cpp
chore: Pump `llama.cpp` version
2 parents 1b10c47 + a6e697e commit d97b1ff

File tree

9 files changed

+11
-15
lines changed

9 files changed

+11
-15
lines changed

.github/workflows/build.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -231,7 +231,7 @@ jobs:
231231
run: |
232232
./install_deps.sh
233233
mkdir build && cd build
234-
cmake -DLLAMA_NATIVE=OFF -DLLAMA_CUBLAS=ON -DLLAMA_CUBLAS=ON -DWHISPER_CUBLAS=ON -DNITRO_VERSION=${{ needs.set-nitro-version.outputs.version }} ..
234+
cmake -DLLAMA_NATIVE=OFF -DLLAMA_CUDA=ON -DWHISPER_CUDA=ON -DNITRO_VERSION=${{ needs.set-nitro-version.outputs.version }} ..
235235
make -j $(nproc)
236236
ls -la
237237
@@ -285,7 +285,7 @@ jobs:
285285
run: |
286286
./install_deps.sh
287287
mkdir build && cd build
288-
cmake -DWHISPER_COREML=1 -DNITRO_VERSION=${{ needs.set-nitro-version.outputs.version }} ..
288+
cmake -DLLAMA_METAL_EMBED_LIBRARY=ON -DWHISPER_COREML=1 -DNITRO_VERSION=${{ needs.set-nitro-version.outputs.version }} ..
289289
CC=gcc-8 make -j $(sysctl -n hw.ncpu)
290290
ls -la
291291

CMakeLists.txt

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,13 +26,13 @@ set(CMAKE_EXPORT_COMPILE_COMMANDS ON)
2626
set(CMAKE_PREFIX_PATH ${CMAKE_CURRENT_SOURCE_DIR}/build_deps/_install)
2727
# This is the critical line for installing another package
2828

29-
if(LLAMA_CUBLAS)
29+
if(LLAMA_CUDA)
3030
cmake_minimum_required(VERSION 3.17)
3131

3232
find_package(CUDAToolkit)
3333
if(CUDAToolkit_FOUND)
3434
message(STATUS "cuBLAS found")
35-
add_compile_definitions(GGML_USE_CUBLAS)
35+
add_compile_definitions(GGML_USE_CUDA)
3636
endif()
3737
endif()
3838

audio.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@
66

77
```bash
88
mkdir build && cd build
9-
cmake -DLLAMA_CUBLAS=ON -DWHISPER_CUBLAS=ON ..
9+
cmake -DLLAMA_CUDA=ON -DWHISPER_CUDA=ON ..
1010
make -j$(nproc)
1111
```
1212

@@ -31,7 +31,7 @@ cmake --build build -j --config Release
3131
```
3232
mkdir -p build
3333
cd build
34-
cmake .. -DLLAMA_CUBLAS=ON -DBUILD_SHARED_LIBS=ON -DWHISPER_CUBLAS=ON -DWHISPER_SDL2=ON
34+
cmake .. -DLLAMA_CUDA=ON -DBUILD_SHARED_LIBS=ON -DWHISPER_CUDA=ON -DWHISPER_SDL2=ON
3535
cmake --build . --config Release
3636
```
3737

context/llama_server_context.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1525,7 +1525,7 @@ struct llama_server_context {
15251525
slot.id, slot.params.n_keep, n_left, n_discard);
15261526
llama_kv_cache_seq_rm(ctx, slot.id, slot.params.n_keep + 1,
15271527
slot.params.n_keep + n_discard + 1);
1528-
llama_kv_cache_seq_shift(ctx, slot.id,
1528+
llama_kv_cache_seq_add(ctx, slot.id,
15291529
slot.params.n_keep + 1 + n_discard,
15301530
slot.n_past, -n_discard);
15311531

controllers/llamaCPP.cc

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -630,10 +630,6 @@ bool llamaCPP::LoadModelImpl(std::shared_ptr<Json::Value> jsonBody) {
630630
log_set_target(llama_log_folder + "llama.log");
631631
} // Set folder for llama log
632632
}
633-
#ifdef GGML_USE_CUBLAS
634-
LOG_INFO << "Setting up GGML CUBLAS PARAMS";
635-
params.mul_mat_q = false;
636-
#endif // GGML_USE_CUBLAS
637633
if (params.model_alias == "unknown") {
638634
params.model_alias = params.model;
639635
}

docs/docs/new/build-source.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ Now, let's generate the build files.
6969

7070
```bash
7171
mkdir build && cd build
72-
cmake -DLLAMA_CUBLAS=ON ..
72+
cmake -DLLAMA_CUDA=ON ..
7373
```
7474

7575
## Build the Application

examples/example-docker/cuda.Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ RUN apt-get update && \
1212
./install_deps.sh && \
1313
mkdir build && \
1414
cd build && \
15-
cmake .. -DDEBUG=ON -DLLAMA_CUBLAS=ON -DLLAMA_CUDA_F16=ON -DLLAMA_CUDA_DMMV_X=64 -DLLAMA_CUDA_MMV_Y=32 && \
15+
cmake .. -DDEBUG=ON -DLLAMA_CUDA=ON -DLLAMA_CUDA_F16=ON -DLLAMA_CUDA_DMMV_X=64 -DLLAMA_CUDA_MMV_Y=32 && \
1616
cmake --build . --config Release -j $(nproc) && \
1717
apt-get remove --purge -y git cmake && \
1818
apt-get autoremove -y && \

llama.cpp

whisper.cpp

0 commit comments

Comments
 (0)