Effortlessly fine-tune your model with Friendli Dedicated Endpoints, which leverages the Parameter-Efficient Fine-Tuning (PEFT) method to reduce training costs while preserving model quality, similar to full-parameter fine-tuning.
.jsonl
or .parquet
dataset file or browse for them on your computer.
If your files meet the required criteria, the blue ‘Upload’ button will be activated, allowing you to complete the process.
tokenizer_config.json
of mistralai/Mistral-7B-Instruct-v0.2
is a template that repeats the messages of a user and an assistant.
Concretely, each row in the “messages” field should follow a format like: [{"role": "user", "content": "The 1st user's message"}, {"role": "assistant", "content": "The 1st assistant's message"}]
.
In this case, HuggingFaceH4/ultrachat_200k
is a dataset that is compatible with the chat template.Learning rate
: Initial learning rate for AdamW optimizer.Batch size
: Total training batch size.Number of training epochs
or Training steps
.
Number of training epochs
: Total number of training epochs.Training steps
: Total number of training steps.Evaluation steps
: Number of steps between model evaluation using the validation dataset.LoRA rank
: The rank of the LoRA parameters (optional).LoRA alpha
: Scaling factor that determines the influence of the low-rank matrices during fine-tuning (optional).LoRA dropout
: Dropout rate applied during fine-tuning (optional).