Skip to content

Commit 238ad3e

Browse files
committed
cmake : do not hide GGML options
ggml-ci
1 parent 0abc6a2 commit 238ad3e

File tree

2 files changed

+13
-4
lines changed

2 files changed

+13
-4
lines changed

CMakeLists.txt

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -82,11 +82,11 @@ set(GGML_FATAL_WARNINGS ${LLAMA_FATAL_WARNINGS})
8282

8383
# change the default for these ggml options
8484
if (NOT DEFINED GGML_LLAMAFILE)
85-
set(GGML_LLAMAFILE ON)
85+
set(GGML_LLAMAFILE_DEFAULT ON)
8686
endif()
8787

8888
if (NOT DEFINED GGML_CUDA_USE_GRAPHS)
89-
set(GGML_CUDA_USE_GRAPHS ON)
89+
set(GGML_CUDA_USE_GRAPHS_DEFAULT ON)
9090
endif()
9191

9292
# transition helpers

ggml/CMakeLists.txt

Lines changed: 11 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -56,6 +56,15 @@ else()
5656
set(GGML_NATIVE_DEFAULT ON)
5757
endif()
5858

59+
# defaults
60+
if (NOT GGML_LLAMAFILE_DEFAULT)
61+
set(GGML_LLAMAFILE_DEFAULT OFF)
62+
endif()
63+
64+
if (NOT GGML_CUDA_USE_GRAPHS_DEFAULT)
65+
set(GGML_CUDA_USE_GRAPHS_DEFAULT OFF)
66+
endif()
67+
5968
# general
6069
option(GGML_STATIC "ggml: static link libraries" OFF)
6170
option(GGML_NATIVE "ggml: enable -march=native flag" ${GGML_NATIVE_DEFAULT})
@@ -110,7 +119,7 @@ option(GGML_ACCELERATE "ggml: enable Accelerate framework"
110119
option(GGML_BLAS "ggml: use BLAS" ${GGML_BLAS_DEFAULT})
111120
set(GGML_BLAS_VENDOR ${GGML_BLAS_VENDOR_DEFAULT} CACHE STRING
112121
"ggml: BLAS library vendor")
113-
option(GGML_LLAMAFILE "ggml: use LLAMAFILE" OFF)
122+
option(GGML_LLAMAFILE "ggml: use LLAMAFILE" ${GGML_LLAMAFILE_DEFAULT})
114123

115124
option(GGML_CUDA "ggml: use CUDA" OFF)
116125
option(GGML_MUSA "ggml: use MUSA" OFF)
@@ -127,7 +136,7 @@ set (GGML_CUDA_PEER_MAX_BATCH_SIZE "128" CACHE STRING
127136
option(GGML_CUDA_NO_PEER_COPY "ggml: do not use peer to peer copies" OFF)
128137
option(GGML_CUDA_NO_VMM "ggml: do not try to use CUDA VMM" OFF)
129138
option(GGML_CUDA_FA_ALL_QUANTS "ggml: compile all quants for FlashAttention" OFF)
130-
option(GGML_CUDA_USE_GRAPHS "ggml: use CUDA graphs (llama.cpp only)" OFF)
139+
option(GGML_CUDA_USE_GRAPHS "ggml: use CUDA graphs (llama.cpp only)" ${GGML_CUDA_USE_GRAPHS_DEFAULT})
131140

132141
option(GGML_HIPBLAS "ggml: use hipBLAS" OFF)
133142
option(GGML_HIP_UMA "ggml: use HIP unified memory architecture" OFF)

0 commit comments

Comments
 (0)