Text Generation
GGUF
chat
Edit model card

QuantFactory/StarDust-12b-v1-GGUF

This is quantized version of Luni/StarDust-12b-v1 created using llama.cpp

Original Model Card

image/png

StarDust-12b-v1

Quants

Description | Usecase

The result of this merge is in my opinion a more vibrant and less generic sonnet inspired prose, it's able to be gentle and harsh where asked. I've personally been trying to get a more spice while also compensating for the Magnum-v2.5 having the issue on my end that it simply won't stop yapping.

  • This model is intended to be used as a Role-playing model.
  • Its direct conversational output is... I can't even say it's luck, it's just not made for it.
  • Extension to Conversational output: The Model is designed for roleplay, direct instructing or general purpose is NOT recommended.

Initial Feedback

Initial feedback shows that the model has a tendency to promote flirting. If this becomes too much try to steer the model with a system prompt to focus on SFW and on-flirty interactions.

Prompting

Edit: ChatML has proven to be the BEST choice.

Both Mistral and ChatML should work though I had better results with ChatML: ChatML Example:

"""<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
<|im_start|>user
Can I ask a question?<|im_end|>
<|im_start|>assistant
"""

Merge Details

Merge Method

This model was merged using the DARE TIES merge method using Sao10K/MN-12B-Lyra-v3 as a base.

Models Merged

The following models were included in the merge:

Special Thanks

Special thanks to the SillyTilly and myself for helping me find the energy to finish this.

Downloads last month
481
GGUF
Model size
12.2B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Examples
Inference API (serverless) is not available, repository is disabled.

Model tree for QuantFactory/StarDust-12b-v1-GGUF

Quantized
this model