Skip to content

Commit 328ded3

Browse files
committed
docs : remove obsolete make references, scripts, examples
ggml-ci
1 parent c536c07 commit 328ded3

File tree

8 files changed

+1
-726
lines changed

8 files changed

+1
-726
lines changed

docs/backend/BLIS.md

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -27,13 +27,6 @@ We recommend using openmp since it's easier to modify the cores being used.
2727

2828
### llama.cpp compilation
2929

30-
Makefile:
31-
32-
```bash
33-
make GGML_BLIS=1 -j
34-
# make GGML_BLIS=1 llama-benchmark-matmult
35-
```
36-
3730
CMake:
3831

3932
```bash

docs/build.md

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,6 @@ In order to build llama.cpp you have four different options.
1818

1919
**Notes**:
2020

21-
- For `Q4_0_4_4` quantization type build, add the `-DGGML_LLAMAFILE=OFF` cmake option. For example, use `cmake -B build -DGGML_LLAMAFILE=OFF`.
2221
- For faster compilation, add the `-j` argument to run multiple jobs in parallel. For example, `cmake --build build --config Release -j 8` will run 8 jobs in parallel.
2322
- For faster repeated compilation, install [ccache](https://ccache.dev/).
2423
- For debug builds, there are two cases:
@@ -337,9 +336,3 @@ For detailed info, such as model/device supports, CANN install, please refer to
337336
### Android
338337

339338
To read documentation for how to build on Android, [click here](./android.md)
340-
341-
### Arm CPU optimized mulmat kernels
342-
343-
Llama.cpp includes a set of optimized mulmat kernels for the Arm architecture, leveraging Arm® Neon™, int8mm and SVE instructions. These kernels are enabled at build time through the appropriate compiler cpu-type flags, such as `-DCMAKE_C_FLAGS=-march=armv8.2a+i8mm+sve`. Note that these optimized kernels require the model to be quantized into one of the formats: `Q4_0_4_4` (Arm Neon), `Q4_0_4_8` (int8mm) or `Q4_0_8_8` (SVE). The SVE mulmat kernel specifically requires a vector width of 256 bits. When running on devices with a different vector width, it is recommended to use the `Q4_0_4_8` (int8mm) or `Q4_0_4_4` (Arm Neon) formats for better performance. Refer to [examples/quantize/README.md](../examples/quantize/README.md) for more information on the quantization formats.
344-
345-
To support `Q4_0_4_4`, you must build with `GGML_NO_LLAMAFILE=1` (`make`) or `-DGGML_LLAMAFILE=OFF` (`cmake`).

examples/base-translate.sh

Lines changed: 0 additions & 61 deletions
This file was deleted.

examples/convert-llama2c-to-ggml/README.md

Lines changed: 1 addition & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -2,11 +2,8 @@
22

33
This example reads weights from project [llama2.c](https://github.com/karpathy/llama2.c) and saves them in ggml compatible format. The vocab that is available in `models/ggml-vocab.bin` is used by default.
44

5-
To convert the model first download the models from the [llama2.c](https://github.com/karpathy/llama2.c) repository:
5+
To convert the model first download the models from the [llama2.c](https://github.com/karpathy/llama2.c) repository.
66

7-
`$ make -j`
8-
9-
After successful compilation, following usage options are available:
107
```
118
usage: ./llama-convert-llama2c-to-ggml [options]
129

examples/imatrix/README.md

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,8 +25,6 @@ For faster computation, make sure to use GPU offloading via the `-ngl` argument
2525
## Example
2626

2727
```bash
28-
GGML_CUDA=1 make -j
29-
3028
# generate importance matrix (imatrix.dat)
3129
./llama-imatrix -m ggml-model-f16.gguf -f train-data.txt -ngl 99
3230

examples/server/README.md

Lines changed: 0 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -188,12 +188,6 @@ services:
188188
189189
`llama-server` is built alongside everything else from the root of the project
190190

191-
- Using `make`:
192-
193-
```bash
194-
make llama-server
195-
```
196-
197191
- Using `CMake`:
198192

199193
```bash
@@ -207,15 +201,6 @@ services:
207201

208202
`llama-server` can also be built with SSL support using OpenSSL 3
209203

210-
- Using `make`:
211-
212-
```bash
213-
# NOTE: For non-system openssl, use the following:
214-
# CXXFLAGS="-I /path/to/openssl/include"
215-
# LDFLAGS="-L /path/to/openssl/lib"
216-
make LLAMA_SERVER_SSL=true llama-server
217-
```
218-
219204
- Using `CMake`:
220205

221206
```bash

scripts/pod-llama.sh

Lines changed: 0 additions & 212 deletions
This file was deleted.

0 commit comments

Comments
 (0)