--- language: - en tags: - not-for-all-audiences --- # Daybreak-Mixtral-8x7b v24.02-7 An experimental model trained on a (currently) private ERP dataset of highly curated niche content (`crestfall/daybreak` as of 2024-02-10). Not suitable for any audience. Model was finetuned on top of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1), and follows that model's instruction format. ## Prompt format: The model uses the Mixtral-8x7b-instruct format (see the base model), but users have repored that Alpaca format gives better results. Try which works for you. ## Training details: The model was trained for 1.83 epochs (eval minima based on 1% of dataset) using Axolotl. See [axolotl.yml](https://huggingface.co/crestf411/crestfall-mixtral-8x7b-hf/blob/main/axolotl/axolotl.yml) for details.