@@ -129,14 +129,14 @@ pub fn build(b: *std.build.Builder) !void {
129
129
const clip = make .obj ("clip" , "examples/llava/clip.cpp" );
130
130
const llava = make .obj ("llava" , "examples/llava/llava.cpp" );
131
131
132
- _ = make .exe ("main" , "examples/main/main.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo , sampling , console , grammar_parser });
133
- _ = make .exe ("quantize" , "examples/quantize/quantize.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo });
134
- _ = make .exe ("perplexity" , "examples/perplexity/perplexity.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo });
135
- _ = make .exe ("embedding" , "examples/embedding/embedding.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo });
136
- _ = make .exe ("finetune" , "examples/finetune/finetune.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo , train });
132
+ _ = make .exe ("main" , "examples/main/main.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo , console , grammar_parser });
133
+ _ = make .exe ("quantize" , "examples/quantize/quantize.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo });
134
+ _ = make .exe ("perplexity" , "examples/perplexity/perplexity.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo });
135
+ _ = make .exe ("embedding" , "examples/embedding/embedding.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo });
136
+ _ = make .exe ("finetune" , "examples/finetune/finetune.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo , train });
137
137
_ = make .exe ("train-text-from-scratch" , "examples/train-text-from-scratch/train-text-from-scratch.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo , train });
138
138
139
- const server = make .exe ("server" , "examples/server/server.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , json_schema_to_grammar , buildinfo , sampling , grammar_parser , clip , llava });
139
+ const server = make .exe ("server" , "examples/server/server.cpp" , &.{ ggml , sgemm , ggml_alloc , ggml_backend , ggml_quants , llama , unicode , unicode_data , common , sampling , json_schema_to_grammar , buildinfo , grammar_parser , clip , llava });
140
140
if (server .target .isWindows ()) {
141
141
server .linkSystemLibrary ("ws2_32" );
142
142
}
0 commit comments