Skip to content

Commit 8d8f065

Browse files
committed
Merge branch 'master' into compilade/mamba2
2 parents 3bc7103 + 401558b commit 8d8f065

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

56 files changed

+21690
-26027
lines changed

CMakePresets.json

Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -48,10 +48,23 @@
4848
}
4949
},
5050

51+
{
52+
"name": "arm64-apple-clang", "hidden": true,
53+
"architecture": { "value": "arm64", "strategy": "external" },
54+
"toolset": { "value": "host=x64", "strategy": "external" },
55+
"cacheVariables": {
56+
"CMAKE_TOOLCHAIN_FILE": "${sourceDir}/cmake/arm64-apple-clang.cmake"
57+
}
58+
},
59+
5160
{ "name": "arm64-windows-llvm-debug" , "inherits": [ "base", "arm64-windows-llvm", "debug" ] },
5261
{ "name": "arm64-windows-llvm-release", "inherits": [ "base", "arm64-windows-llvm", "reldbg" ] },
5362
{ "name": "arm64-windows-llvm+static-release", "inherits": [ "base", "arm64-windows-llvm", "reldbg", "static" ] },
5463

64+
{ "name": "arm64-apple-clang-debug" , "inherits": [ "base", "arm64-apple-clang", "debug" ] },
65+
{ "name": "arm64-apple-clang-release" , "inherits": [ "base", "arm64-apple-clang", "reldbg" ] },
66+
{ "name": "arm64-apple-clang+static-release" , "inherits": [ "base", "arm64-apple-clang", "reldbg", "static" ] },
67+
5568
{ "name": "arm64-windows-msvc-debug" , "inherits": [ "base", "arm64-windows-msvc", "debug" ] },
5669
{ "name": "arm64-windows-msvc-release", "inherits": [ "base", "arm64-windows-msvc", "reldbg" ] },
5770
{ "name": "arm64-windows-msvc+static-release", "inherits": [ "base", "arm64-windows-msvc", "reldbg", "static" ] },

Makefile

Lines changed: 14 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
# Define the default target now so that it is always the first target
22
BUILD_TARGETS = \
33
libllava.a \
4-
llama-baby-llama \
54
llama-batched \
65
llama-batched-bench \
76
llama-bench \
@@ -34,6 +33,7 @@ BUILD_TARGETS = \
3433
llama-save-load-state \
3534
llama-server \
3635
llama-simple \
36+
llama-simple-chat \
3737
llama-speculative \
3838
llama-tokenize \
3939
llama-vdot \
@@ -55,14 +55,14 @@ TEST_TARGETS = \
5555
tests/test-llama-grammar \
5656
tests/test-log \
5757
tests/test-model-load-cancel \
58-
tests/test-opt \
5958
tests/test-quantize-fns \
6059
tests/test-quantize-perf \
6160
tests/test-rope \
6261
tests/test-sampling \
6362
tests/test-tokenizer-0 \
6463
tests/test-tokenizer-1-bpe \
6564
tests/test-tokenizer-1-spm
65+
# tests/test-opt \
6666
6767
# Legacy build targets that were renamed in #7809, but should still be removed when the project is cleaned
6868
LEGACY_TARGETS_CLEAN = main quantize quantize-stats perplexity imatrix embedding vdot q8dot convert-llama2c-to-ggml \
@@ -915,6 +915,7 @@ endif # GGML_METAL
915915

916916
OBJ_GGML += \
917917
ggml/src/ggml.o \
918+
ggml/src/ggml-cpu.o \
918919
ggml/src/ggml-alloc.o \
919920
ggml/src/ggml-backend.o \
920921
ggml/src/ggml-quants.o \
@@ -935,7 +936,6 @@ OBJ_COMMON = \
935936
common/console.o \
936937
common/ngram-cache.o \
937938
common/sampling.o \
938-
common/train.o \
939939
common/build-info.o \
940940
common/json-schema-to-grammar.o
941941

@@ -1047,6 +1047,12 @@ ggml/src/ggml.o: \
10471047
ggml/include/ggml.h
10481048
$(CC) $(CFLAGS) -c $< -o $@
10491049

1050+
ggml/src/ggml-cpu.o: \
1051+
ggml/src/ggml-cpu.c \
1052+
ggml/include/ggml.h \
1053+
ggml/src/ggml-common.h
1054+
$(CC) $(CFLAGS) -c $< -o $@
1055+
10501056
ggml/src/ggml-alloc.o: \
10511057
ggml/src/ggml-alloc.c \
10521058
ggml/include/ggml.h \
@@ -1212,11 +1218,6 @@ common/json-schema-to-grammar.o: \
12121218
common/json-schema-to-grammar.h
12131219
$(CXX) $(CXXFLAGS) -c $< -o $@
12141220

1215-
common/train.o: \
1216-
common/train.cpp \
1217-
common/train.h
1218-
$(CXX) $(CXXFLAGS) -c $< -o $@
1219-
12201221
common/ngram-cache.o: \
12211222
common/ngram-cache.cpp \
12221223
common/ngram-cache.h
@@ -1287,6 +1288,11 @@ llama-simple: examples/simple/simple.cpp \
12871288
$(CXX) $(CXXFLAGS) -c $< -o $(call GET_OBJ_FILE, $<)
12881289
$(CXX) $(CXXFLAGS) $(filter-out %.h $<,$^) $(call GET_OBJ_FILE, $<) -o $@ $(LDFLAGS)
12891290

1291+
llama-simple-chat: examples/simple-chat/simple-chat.cpp \
1292+
$(OBJ_ALL)
1293+
$(CXX) $(CXXFLAGS) -c $< -o $(call GET_OBJ_FILE, $<)
1294+
$(CXX) $(CXXFLAGS) $(filter-out %.h $<,$^) $(call GET_OBJ_FILE, $<) -o $@ $(LDFLAGS)
1295+
12901296
llama-tokenize: examples/tokenize/tokenize.cpp \
12911297
$(OBJ_ALL)
12921298
$(CXX) $(CXXFLAGS) -c $< -o $(call GET_OBJ_FILE, $<)
@@ -1384,11 +1390,6 @@ llama-bench: examples/llama-bench/llama-bench.cpp \
13841390
$(CXX) $(CXXFLAGS) -c $< -o $(call GET_OBJ_FILE, $<)
13851391
$(CXX) $(CXXFLAGS) $(filter-out %.h $<,$^) $(call GET_OBJ_FILE, $<) -o $@ $(LDFLAGS)
13861392

1387-
llama-baby-llama: examples/baby-llama/baby-llama.cpp \
1388-
$(OBJ_ALL)
1389-
$(CXX) $(CXXFLAGS) -c $< -o $(call GET_OBJ_FILE, $<)
1390-
$(CXX) $(CXXFLAGS) $(filter-out %.h $<,$^) $(call GET_OBJ_FILE, $<) -o $@ $(LDFLAGS)
1391-
13921393
llama-export-lora: examples/export-lora/export-lora.cpp \
13931394
$(OBJ_ALL)
13941395
$(CXX) $(CXXFLAGS) -c $< -o $(call GET_OBJ_FILE, $<)

Package.swift

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,7 @@ var sources = [
1010
"src/unicode.cpp",
1111
"src/unicode-data.cpp",
1212
"ggml/src/ggml.c",
13+
"ggml/src/ggml-cpu.c",
1314
"ggml/src/ggml-alloc.c",
1415
"ggml/src/ggml-backend.cpp",
1516
"ggml/src/ggml-quants.c",

README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,8 @@ Inference of Meta's [LLaMA](https://arxiv.org/abs/2302.13971) model (and others)
1717

1818
## Hot topics
1919

20-
- **Hugging Face Inference Endpoints now support GGUF out of the box! https://github.com/ggerganov/llama.cpp/discussions/9669**
20+
- **Introducing GGUF-my-LoRA** https://github.com/ggerganov/llama.cpp/discussions/10123
21+
- Hugging Face Inference Endpoints now support GGUF out of the box! https://github.com/ggerganov/llama.cpp/discussions/9669
2122
- Hugging Face GGUF editor: [discussion](https://github.com/ggerganov/llama.cpp/discussions/9268) | [tool](https://huggingface.co/spaces/CISCai/gguf-editor)
2223

2324
----

0 commit comments

Comments
 (0)