File tree 1 file changed +9
-1
lines changed
1 file changed +9
-1
lines changed Original file line number Diff line number Diff line change @@ -13,9 +13,13 @@ elif [[ "$arg1" == '--quantize' || "$arg1" == '-q' ]]; then
13
13
exec ./llama-quantize " $@ "
14
14
elif [[ " $arg1 " == ' --run' || " $arg1 " == ' -r' ]]; then
15
15
exec ./llama-cli " $@ "
16
+ elif [[ " $arg1 " == ' --bench' || " $arg1 " == ' -b' ]]; then
17
+ exec ./llama-bench " $@ "
18
+ elif [[ " $arg1 " == ' --perplexity' || " $arg1 " == ' -p' ]]; then
19
+ exec ./llama-perplexity " $@ "
16
20
elif [[ " $arg1 " == ' --all-in-one' || " $arg1 " == ' -a' ]]; then
17
21
echo " Converting PTH to GGML..."
18
- for i in ` ls $1 /$2 /ggml-model-f16.bin* ` ; do
22
+ for i in $( ls $1 /$2 /ggml-model-f16.bin* ) ; do
19
23
if [ -f " ${i/ f16/ q4_0} " ]; then
20
24
echo " Skip model quantization, it already exists: ${i/ f16/ q4_0} "
21
25
else
30
34
echo " Available commands: "
31
35
echo " --run (-r): Run a model previously converted into ggml"
32
36
echo " ex: -m /models/7B/ggml-model-q4_0.bin -p \" Building a website can be done in 10 simple steps:\" -n 512"
37
+ echo " --bench (-b): Benchmark the performance of the inference for various parameters."
38
+ echo " ex: -m model.gguf"
39
+ echo " --perplexity (-p): Measure the perplexity of a model over a given text."
40
+ echo " ex: -m model.gguf -f file.txt"
33
41
echo " --convert (-c): Convert a llama model into ggml"
34
42
echo " ex: --outtype f16 \" /models/7B/\" "
35
43
echo " --quantize (-q): Optimize with quantization process ggml"
You can’t perform that action at this time.
0 commit comments