This model was trained as part of a series of experiments testing the performance of pure DPO vs SFT vs ORPO, all supported by Unsloth/Huggingface TRL.
Benchmarks
Average 56.93
ARC 56.83
HellaSwag 79.75
MMLU 56.76
TruthfulQA 46.29
Winogrande 76.64
GSM8K 25.32
Training Details
Duration: ~6-8 hours on one Kaggle T4 with Unsloth
Model: https://huggingface.co/unsloth/mistral-7b-v0.2-bnb-4bit
Dataset: https://huggingface.co/datasets/argilla/dpo-mix-7k
Rank: 8
Alpha: 16
Learning rate: 5e-4
Batch size: 8
Epochs: 1
Learning rate scheduler: Linear
Prompt Format: ChatML
<|im_start|>system
You are a helpful assistant.<|im_end|>
<|im_start|>user
Why is the sky blue?<|im_end|>
<|im_start|>assistant
WanDB Reports