@@ -128,17 +128,18 @@ pub fn build(b: *std.build.Builder) !void {
128
128
const llama = make .obj ("llama" , "llama.cpp" );
129
129
const common = make .obj ("common" , "common/common.cpp" );
130
130
const console = make .obj ("console" , "common/console.cpp" );
131
+ const sampling = make .obj ("sampling" , "common/sampling.cpp" );
131
132
const grammar_parser = make .obj ("grammar-parser" , "common/grammar-parser.cpp" );
132
133
const train = make .obj ("train" , "common/train.cpp" );
133
134
134
- _ = make .exe ("main" , "examples/main/main.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , console , grammar_parser });
135
+ _ = make .exe ("main" , "examples/main/main.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , sampling , console , grammar_parser });
135
136
_ = make .exe ("quantize" , "examples/quantize/quantize.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common });
136
137
_ = make .exe ("perplexity" , "examples/perplexity/perplexity.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common });
137
138
_ = make .exe ("embedding" , "examples/embedding/embedding.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common });
138
139
_ = make .exe ("finetune" , "examples/finetune/finetune.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , train });
139
140
_ = make .exe ("train-text-from-scratch" , "examples/train-text-from-scratch/train-text-from-scratch.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , train });
140
141
141
- const server = make .exe ("server" , "examples/server/server.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , grammar_parser });
142
+ const server = make .exe ("server" , "examples/server/server.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , sampling , grammar_parser });
142
143
if (server .target .isWindows ()) {
143
144
server .linkSystemLibrary ("ws2_32" );
144
145
}
0 commit comments