You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to do a full fine-tune of TinyLLama on a Mac M1 with 8 GB of RAM.
Fine-tuning Script:
#!/bin/bash
# Path to the llama.cpp directory
LLAMA_CPP_DIR="./llama.cpp"
# Train the model
echo "Starting training process..."
$LLAMA_CPP_DIR/train-text-from-scratch \
--vocab-model $LLAMA_CPP_DIR/models/ggml-vocab-llama.gguf \
--ctx 4096 --embd 2048 --head 32 --layer 22 \
--checkpoint-in $LLAMA_CPP_DIR/models/ggml-model-f32.gguf \
--checkpoint-out $LLAMA_CPP_DIR/models/chk-tinyllama-1.1b-3t-chat-LATEST.gguf \
--model-out $LLAMA_CPP_DIR/models/tinyllama-1.1b-intermediate-step-1431k-3t-chat.gguf \
--train-data "data/train_data.txt" \
-t 6 -b 1 --seed 1 --adam-iter 256 \
# --no-checkpointing
# Check if the train command succeeded
if [ $? -ne 0 ]; then
echo "Training process failed."
exit 1
fi
# Run prediction with the fine-tuned model
echo "Running prediction with the fine-tuned model..."
$LLAMA_CPP_DIR/main -m $LLAMA_CPP_DIR/models/tinyllama-1.1b-intermediate-step-1431k-3t-chat.gguf
# End of script
echo "Fine-tuning process completed."
It's not clear from the examples folder, but it seems that one needs to use an 'F32' model, so I prepared an F32 model using:
Note that I want to train on context of 4096, so I changed the config.json to have max embeddings of 4096 after downloading config.json and pytorch_model.bin from the TinyLlama repo.
The model is clearly not being loaded correctly. Is there a specific vocab file that I should be using. As per the logs, you can see that the llama vocab file is being used but it seems that values for the layers are not being correctly overwritten. Also, there are 4 kv groups, which is different than Llama 2 so clearly I need to input that info somehow.
Is it true that I can only input an f32 model for a full fine tune? (it would be great if I could input any gguf quant)?
The text was updated successfully, but these errors were encountered:
I am trying to do a full fine-tune of TinyLLama on a Mac M1 with 8 GB of RAM.
Fine-tuning Script:
It's not clear from the examples folder, but it seems that one needs to use an 'F32' model, so I prepared an F32 model using:
Note that I want to train on context of 4096, so I changed the config.json to have max embeddings of 4096 after downloading config.json and pytorch_model.bin from the TinyLlama repo.
When I run './fine-tune.sh' I get:
Summary of Issues:
The text was updated successfully, but these errors were encountered: