Edit model card

Noromaid-v0.1-mixtral-8x7b-v3 4bpw

Exllama quant of NeverSleep/Noromaid-v0.1-mixtral-8x7b-v3

You probably want to use the 3.5bpw version, it lands very close to 24gb of vram with 8bit cache enabled. GGUF is probably preferable if you have less than 24gb.

Other BPW's 2.7bpw, 3.0bpw, 3.5bpw, 4.0bpw

Prompt format:

### Instruction:
{system prompt}

### Input:
{input}

### Response:
{reply}

Sillytavern presets

Discussion with presets

Contact

Kooten on discord.

Downloads last month
12
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.