Skip to content

Commit 02f92f6

Browse files
committed
Merge branch 'upstream' into concedo_experimental
# Conflicts: # .devops/full-cuda.Dockerfile # .devops/full-rocm.Dockerfile # .devops/llama-cli-cuda.Dockerfile # .devops/llama-cli-rocm.Dockerfile # .devops/llama-cli-vulkan.Dockerfile # .devops/llama-cpp-cuda.srpm.spec # .devops/llama-server-cuda.Dockerfile # .devops/llama-server-rocm.Dockerfile # .devops/llama-server-vulkan.Dockerfile # .github/workflows/build.yml # .github/workflows/docker.yml # CMakeLists.txt # Makefile # README.md # examples/llama.android/llama/src/main/cpp/CMakeLists.txt # flake.lock # ggml/CMakeLists.txt # ggml/src/CMakeLists.txt # grammars/README.md # scripts/sync-ggml-am.sh # scripts/sync-ggml.last # tests/test-chat-template.cpp # tests/test-grammar-integration.cpp # tests/test-json-schema-to-grammar.cpp
2 parents 8421243 + 72272b8 commit 02f92f6

22 files changed

+628
-178
lines changed

.devops/llama-cli-intel.Dockerfile

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -2,19 +2,19 @@ ARG ONEAPI_VERSION=2024.1.1-devel-ubuntu22.04
22

33
FROM intel/oneapi-basekit:$ONEAPI_VERSION as build
44

5-
ARG LLAMA_SYCL_F16=OFF
5+
ARG GGML_SYCL_F16=OFF
66
RUN apt-get update && \
77
apt-get install -y git
88

99
WORKDIR /app
1010

1111
COPY . .
1212

13-
RUN if [ "${LLAMA_SYCL_F16}" = "ON" ]; then \
14-
echo "LLAMA_SYCL_F16 is set" && \
15-
export OPT_SYCL_F16="-DLLAMA_SYCL_F16=ON"; \
13+
RUN if [ "${GGML_SYCL_F16}" = "ON" ]; then \
14+
echo "GGML_SYCL_F16 is set" && \
15+
export OPT_SYCL_F16="-DGGML_SYCL_F16=ON"; \
1616
fi && \
17-
cmake -B build -DLLAMA_SYCL=ON -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx ${OPT_SYCL_F16} && \
17+
cmake -B build -DGGML_SYCL=ON -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx ${OPT_SYCL_F16} && \
1818
cmake --build build --config Release --target llama-cli
1919

2020
FROM intel/oneapi-basekit:$ONEAPI_VERSION as runtime

.devops/llama-server-intel.Dockerfile

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -2,19 +2,19 @@ ARG ONEAPI_VERSION=2024.1.1-devel-ubuntu22.04
22

33
FROM intel/oneapi-basekit:$ONEAPI_VERSION as build
44

5-
ARG LLAMA_SYCL_F16=OFF
5+
ARG GGML_SYCL_F16=OFF
66
RUN apt-get update && \
77
apt-get install -y git libcurl4-openssl-dev
88

99
WORKDIR /app
1010

1111
COPY . .
1212

13-
RUN if [ "${LLAMA_SYCL_F16}" = "ON" ]; then \
14-
echo "LLAMA_SYCL_F16 is set" && \
15-
export OPT_SYCL_F16="-DLLAMA_SYCL_F16=ON"; \
13+
RUN if [ "${GGML_SYCL_F16}" = "ON" ]; then \
14+
echo "GGML_SYCL_F16 is set" && \
15+
export OPT_SYCL_F16="-DGGML_SYCL_F16=ON"; \
1616
fi && \
17-
cmake -B build -DLLAMA_SYCL=ON -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx -DLLAMA_CURL=ON ${OPT_SYCL_F16} && \
17+
cmake -B build -DGGML_SYCL=ON -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx -DLLAMA_CURL=ON ${OPT_SYCL_F16} && \
1818
cmake --build build --config Release --target llama-server
1919

2020
FROM intel/oneapi-basekit:$ONEAPI_VERSION as runtime

0 commit comments

Comments
 (0)