Noromaid-v0.1-mixtral-8x7b-v3 4bpw
Exllama quant of NeverSleep/Noromaid-v0.1-mixtral-8x7b-v3
You probably want to use the 3.5bpw version, it lands very close to 24gb of vram with 8bit cache enabled. GGUF is probably preferable if you have less than 24gb.
Other BPW's 2.7bpw, 3.0bpw, 3.5bpw, 4.0bpw
Prompt format:
### Instruction:
{system prompt}
### Input:
{input}
### Response:
{reply}
Sillytavern presets
- Context : https://files.catbox.moe/vnnsra.json
- Instruct : https://files.catbox.moe/7sudj5.json
- Preset : https://files.catbox.moe/og94e5.json
Contact
Kooten on discord.
- Downloads last month
- 12
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.