File size: 1,272 Bytes
926b2ba f73df71 926b2ba f73df71 3da4a49 f73df71 3da4a49 f73df71 1d08343 e5d113f 1d08343 f73df71 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 |
---
library_name: transformers
license: apache-2.0
base_model: allenai/OLMoE-1B-7B-0924
tags:
- axolotl
- moe
- roleplay
model-index:
- name: MoE-girl_1BA_7BT
results: []
---
# MoE Girl 1bA 7bT
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/634262af8d8089ebaefd410e/kTXXSSSqpb21rfyOX7FUa.jpeg)
A finetune of OLMoE by AllenAI designed for roleplaying (and maybe general usecases if you try hard enough).
## Disclaimer
PLEASE do not expect godliness out of this, it's a model with 1 billion active parameters. Expect something more akin to Gemma 2 2B, not Llama 3 8B.
## Quants
GGUF (requires a newish version of llama.cpp or kobold.cpp 1.76):
- [mradermacher's imatrix quants](https://huggingface.co/mradermacher/MoE-Girl-1BA-7BT-i1-GGUF)
- [our static quants](https://huggingface.co/allura-org/MoE-Girl-1BA-7BT-GGUF)
## Prompting
Use ChatML.
```
<|im_start|>system
You are a helpful assistant who talks like a pirate.<|im_end|>
<|im_start|>user
Hello there!<|im_end|>
<|im_start|>assistant
Yarr harr harr, me matey!<|im_end|>
```
## Thanks
Special thanks to the members of Allura for testing and emotional support, as well as the creators of all the datasets that were used in the Special Sauce used to train this model. I love you all <3 - Fizz |