File tree Expand file tree Collapse file tree 3 files changed +8
-8
lines changed Expand file tree Collapse file tree 3 files changed +8
-8
lines changed Original file line number Diff line number Diff line change @@ -133,7 +133,7 @@ The docker build option is currently limited to *intel GPU* targets.
133
133
### Build image
134
134
``` sh
135
135
# Using FP16
136
- docker build -t llama-cpp-sycl --build-arg=" GGML_SYCL_F16=ON" -f .devops/llama-cli- intel.Dockerfile .
136
+ docker build -t llama-cpp-sycl --build-arg=" GGML_SYCL_F16=ON" --target light - f .devops/intel.Dockerfile .
137
137
```
138
138
139
139
* Notes* :
Original file line number Diff line number Diff line change @@ -286,7 +286,7 @@ You don't need to install Vulkan SDK. It will be installed inside the container.
286
286
287
287
``` sh
288
288
# Build the image
289
- docker build -t llama-cpp-vulkan -f .devops/llama-cli- vulkan.Dockerfile .
289
+ docker build -t llama-cpp-vulkan --target light - f .devops/vulkan.Dockerfile .
290
290
291
291
# Then, use it:
292
292
docker run -it --rm -v " $( pwd) :/app:Z" --device /dev/dri/renderD128:/dev/dri/renderD128 --device /dev/dri/card1:/dev/dri/card1 llama-cpp-vulkan -m " /app/models/YOUR_MODEL_FILE" -p " Building a website can be done in 10 simple steps:" -n 400 -e -ngl 33
Original file line number Diff line number Diff line change @@ -60,9 +60,9 @@ Assuming one has the [nvidia-container-toolkit](https://github.com/NVIDIA/nvidia
60
60
## Building Docker locally
61
61
62
62
``` bash
63
- docker build -t local/llama.cpp:full-cuda -f .devops/full- cuda.Dockerfile .
64
- docker build -t local/llama.cpp:light-cuda -f .devops/llama-cli- cuda.Dockerfile .
65
- docker build -t local/llama.cpp:server-cuda -f .devops/llama-server- cuda.Dockerfile .
63
+ docker build -t local/llama.cpp:full-cuda --target full - f .devops/cuda.Dockerfile .
64
+ docker build -t local/llama.cpp:light-cuda --target light - f .devops/cuda.Dockerfile .
65
+ docker build -t local/llama.cpp:server-cuda --target server - f .devops/cuda.Dockerfile .
66
66
```
67
67
68
68
You may want to pass in some different ` ARGS ` , depending on the CUDA environment supported by your container host, as well as the GPU architecture.
@@ -95,9 +95,9 @@ Assuming one has the [mt-container-toolkit](https://developer.mthreads.com/musa/
95
95
## Building Docker locally
96
96
97
97
``` bash
98
- docker build -t local/llama.cpp:full-musa -f .devops/full- musa.Dockerfile .
99
- docker build -t local/llama.cpp:light-musa -f .devops/llama-cli- musa.Dockerfile .
100
- docker build -t local/llama.cpp:server-musa -f .devops/llama-server- musa.Dockerfile .
98
+ docker build -t local/llama.cpp:full-musa --target full - f .devops/musa.Dockerfile .
99
+ docker build -t local/llama.cpp:light-musa --target light - f .devops/musa.Dockerfile .
100
+ docker build -t local/llama.cpp:server-musa --target server - f .devops/musa.Dockerfile .
101
101
```
102
102
103
103
You may want to pass in some different ` ARGS ` , depending on the MUSA environment supported by your container host, as well as the GPU architecture.
You can’t perform that action at this time.
0 commit comments