File tree Expand file tree Collapse file tree 4 files changed +7
-7
lines changed Expand file tree Collapse file tree 4 files changed +7
-7
lines changed Original file line number Diff line number Diff line change @@ -85,8 +85,8 @@ if (NOT DEFINED GGML_LLAMAFILE)
85
85
set (GGML_LLAMAFILE_DEFAULT ON )
86
86
endif ()
87
87
88
- if (NOT DEFINED GGML_CUDA_USE_GRAPHS )
89
- set (GGML_CUDA_USE_GRAPHS_DEFAULT ON )
88
+ if (NOT DEFINED GGML_CUDA_GRAPHS )
89
+ set (GGML_CUDA_GRAPHS_DEFAULT ON )
90
90
endif ()
91
91
92
92
# transition helpers
Original file line number Diff line number Diff line change @@ -619,7 +619,7 @@ ifdef GGML_CUDA
619
619
CUDA_PATH ?= /usr/local/cuda
620
620
endif
621
621
622
- MK_CPPFLAGS += -DGGML_USE_CUDA -I$(CUDA_PATH)/include -I$(CUDA_PATH)/targets/$(UNAME_M)-linux/include -DGGML_CUDA_USE_GRAPHS
622
+ MK_CPPFLAGS += -DGGML_USE_CUDA -DGGML_CUDA_USE_GRAPHS - I$(CUDA_PATH)/include -I$(CUDA_PATH)/targets/$(UNAME_M)-linux/include
623
623
MK_LDFLAGS += -lcuda -lcublas -lculibos -lcudart -lcublasLt -lpthread -ldl -lrt -L$(CUDA_PATH)/lib64 -L/usr/lib64 -L$(CUDA_PATH)/targets/$(UNAME_M)-linux/lib -L$(CUDA_PATH)/lib64/stubs -L/usr/lib/wsl/lib
624
624
MK_NVCCFLAGS += -use_fast_math
625
625
endif # GGML_MUSA
Original file line number Diff line number Diff line change @@ -61,8 +61,8 @@ if (NOT GGML_LLAMAFILE_DEFAULT)
61
61
set (GGML_LLAMAFILE_DEFAULT OFF )
62
62
endif ()
63
63
64
- if (NOT GGML_CUDA_USE_GRAPHS_DEFAULT )
65
- set (GGML_CUDA_USE_GRAPHS_DEFAULT OFF )
64
+ if (NOT GGML_CUDA_GRAPHS_DEFAULT )
65
+ set (GGML_CUDA_GRAPHS_DEFAULT OFF )
66
66
endif ()
67
67
68
68
# general
@@ -136,7 +136,7 @@ set (GGML_CUDA_PEER_MAX_BATCH_SIZE "128" CACHE STRING
136
136
option (GGML_CUDA_NO_PEER_COPY "ggml: do not use peer to peer copies" OFF )
137
137
option (GGML_CUDA_NO_VMM "ggml: do not try to use CUDA VMM" OFF )
138
138
option (GGML_CUDA_FA_ALL_QUANTS "ggml: compile all quants for FlashAttention" OFF )
139
- option (GGML_CUDA_USE_GRAPHS "ggml: use CUDA graphs (llama.cpp only)" ${GGML_CUDA_USE_GRAPHS_DEFAULT } )
139
+ option (GGML_CUDA_GRAPHS "ggml: use CUDA graphs (llama.cpp only)" ${GGML_CUDA_GRAPHS_DEFAULT } )
140
140
141
141
option (GGML_HIPBLAS "ggml: use hipBLAS" OFF )
142
142
option (GGML_HIP_UMA "ggml: use HIP unified memory architecture" OFF )
Original file line number Diff line number Diff line change @@ -329,7 +329,7 @@ if (GGML_CUDA)
329
329
add_compile_definitions (K_QUANTS_PER_ITERATION=${GGML_CUDA_KQUANTS_ITER} )
330
330
add_compile_definitions (GGML_CUDA_PEER_MAX_BATCH_SIZE=${GGML_CUDA_PEER_MAX_BATCH_SIZE} )
331
331
332
- if (GGML_CUDA_USE_GRAPHS )
332
+ if (GGML_CUDA_GRAPHS )
333
333
add_compile_definitions (GGML_CUDA_USE_GRAPHS )
334
334
endif ()
335
335
You can’t perform that action at this time.
0 commit comments