aashish1904 commited on
Commit
95d7e7e
1 Parent(s): 34d8a2a

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +185 -0
README.md ADDED
@@ -0,0 +1,185 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ base_model:
5
+ - nothingiisreal/L3.1-8B-Celeste-V1.5
6
+ - Sao10K/Llama-3.1-8B-Stheno-v3.4
7
+ - Sao10K/L3.1-8B-Niitama-v1.1
8
+ - arcee-ai/Llama-3.1-SuperNova-Lite
9
+ - akjindal53244/Llama-3.1-Storm-8B
10
+ - arcee-ai/Llama-Spark
11
+ - grimjim/Llama-3-Instruct-abliteration-LoRA-8B
12
+ - crestf411/sunfall-peft
13
+ tags:
14
+ - llama
15
+ - merge
16
+ - llama3
17
+ - mixtral
18
+ library_name: transformers
19
+
20
+ ---
21
+
22
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
23
+
24
+
25
+ # QuantFactory/L3.1-Celestial-Stone-2x8B-GGUF
26
+ This is quantized version of [v000000/L3.1-Celestial-Stone-2x8B](https://huggingface.co/v000000/L3.1-Celestial-Stone-2x8B) created using llama.cpp
27
+
28
+ # Original Model Card
29
+
30
+
31
+
32
+ > [!WARNING]
33
+ > **Content:**<br>
34
+ > This models output's can be a bit unhinged.
35
+
36
+ # Llama-3.1-Celestial-Stone-2x8B (BF16)
37
+
38
+ * *Mixture of Experts (14B).*
39
+
40
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64f74b6e6389380c77562762/lBrXRa3sVRinE3cabs-oQ.png)
41
+
42
+ Both experts are used in tandem when generating a token.
43
+
44
+ ------------------------------------------------------------------------------
45
+
46
+ * *Llama.CPP - GGUF.*
47
+
48
+ # Thank you mradermacher for the quants!
49
+
50
+ ----> [GGUF iMatrix](https://huggingface.co/mradermacher/L3.1-Celestial-Stone-2x8B-i1-GGUF)
51
+
52
+ ----> [GGUF static](https://huggingface.co/mradermacher/L3.1-Celestial-Stone-2x8B-GGUF)
53
+
54
+ Other alternative quants:
55
+
56
+ ----> [Q8_0 GGUF by dasChronos1](https://huggingface.co/dasChronos1/L3.1-Celestial-Stone-2x8B-Q8_0-GGUF)
57
+
58
+ ----> [Q6_K GGUF](https://huggingface.co/v000000/L3.1-Celestial-Stone-2x8B-Q6_K-GGUF)
59
+
60
+ ----> [Q4_K_M GGUF by aashish1904](https://huggingface.co/aashish1904/L3.1-Celestial-Stone-2x8B-Q4_K_M-GGUF)
61
+
62
+ ----> [Q2_K GGUF by aashish1904](https://huggingface.co/aashish1904/L3.1-Celestial-Stone-2x8B-Q2_K-GGUF)
63
+
64
+ ------------------------------------------------------------------------------
65
+
66
+ *The first expert* is Instruct 405B distillation/RP vector merge <b>(Supernova-Lite, Niitama1.1, Storm)</b>
67
+
68
+ *The second expert* is ERP/Reddit data merge <b>(Celeste1.5, Stheno3.4, Storm)</b>
69
+
70
+ -------------------------------------------------------------------------------
71
+
72
+ *The base model* is <b>Sao10k/L3.1-Stheno-3.4</b> with the <b>Sunfall LoRa 0.6.1</b> to make it understand SillyTavern prompts and storywriting better.
73
+
74
+ -------------------------------------------------------------------------------
75
+
76
+ # Prompt Template:
77
+ ```bash
78
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
79
+
80
+ {system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
81
+
82
+ {input}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
83
+
84
+ {output}<|eot_id|>
85
+
86
+ ```
87
+
88
+ * *Other Details:*
89
+
90
+ *The model has 131072 context length, and is on Llama-3.1 and Mixtral architecture.*
91
+
92
+ *I did not abliterate the base model at all, so it will refuse zero-shot unethical questions. I recommend avoiding keywords like 'assistant, helpful, kind'*
93
+
94
+ # Recipe (I'm sorry...):
95
+ ```yaml
96
+ slices:
97
+ - sources:
98
+ - model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
99
+ layer_range: [0, 32]
100
+ - model: akjindal53244/Llama-3.1-Storm-8B
101
+ layer_range: [0, 32]
102
+ merge_method: nearswap
103
+ base_model: Sao10K/L3.1-8B-Niitama-v1.1+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
104
+ parameters:
105
+ t:
106
+ - value: 0.0001
107
+ dtype: bfloat16
108
+ out_type: float16
109
+ slices:
110
+ - sources:
111
+ - model: v000000/Llama-3.1-8B-Stheno-v3.4-abliterated
112
+ layer_range: [0, 32]
113
+ - model: akjindal53244/Llama-3.1-Storm-8B
114
+ layer_range: [0, 32]
115
+ merge_method: slerp
116
+ base_model: v000000/Llama-3.1-8B-Stheno-v3.4-abliterated
117
+ parameters:
118
+ t:
119
+ - filter: self_attn
120
+ value: [0.1, 0.6, 0.3, 0.8, 0.5]
121
+ - filter: mlp
122
+ value: [0.9, 0.4, 0.7, 0.2, 0.5]
123
+ - value: 0.5
124
+ dtype: float32
125
+ models:
126
+ - model: arcee-ai/Llama-3.1-SuperNova-Lite
127
+ parameters:
128
+ weight: 1.0
129
+ - model: v000000/L3.1-Niitorm-8B-t0.0001
130
+ parameters:
131
+ weight: 0.4
132
+ merge_method: task_arithmetic
133
+ base_model: arcee-ai/Llama-3.1-SuperNova-Lite
134
+ parameters:
135
+ normalize: false
136
+ dtype: float16
137
+ models:
138
+ - model: arcee-ai/Llama-3.1-SuperNova-Lite
139
+ parameters:
140
+ weight: 0.0
141
+ - model: v000000/L3.1-Niitorm-8B-t0.0001
142
+ parameters:
143
+ weight: 1.25
144
+ merge_method: task_arithmetic
145
+ base_model: arcee-ai/Llama-3.1-SuperNova-Lite
146
+ parameters:
147
+ normalize: false
148
+ dtype: float16
149
+ models:
150
+ - model: v000000/L3.1-8B-RP-Test-003-Task_Arithmetic
151
+ merge_method: slerp
152
+ base_model: v000000/L3.1-8B-RP-Test-002-Task_Arithmetic+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
153
+ parameters:
154
+ t:
155
+ - value: [0, 0, 0.3, 0.4, 0.5, 0.6, 0.5, 0.4, 0.3, 0, 0]
156
+ dtype: float16
157
+ base_model: nothingiisreal/L3.1-8B-Celeste-V1.5+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
158
+ dtype: bfloat16
159
+ merge_method: task_arithmetic
160
+ parameters:
161
+ normalize: false
162
+ slices:
163
+ - sources:
164
+ - layer_range: [0, 32]
165
+ model: nothingiisreal/L3.1-8B-Celeste-V1.5+grimjim/Llama-3-Instruct-abliteration-LoRA-8B
166
+ parameters:
167
+ weight: 0.7
168
+ - layer_range: [0, 32]
169
+ model: v000000/L3.1-Sthenorm-8B
170
+ parameters:
171
+ weight: 0.2
172
+ - layer_range: [0, 32]
173
+ model: nothingiisreal/L3.1-8B-Celeste-V1.5
174
+ parameters:
175
+ weight: 0.2
176
+ base_model: crestf411/L3.1-8B-sunfall-stheno-v0.6.1
177
+ experts_per_token: 2
178
+ local_experts: 2
179
+ gate_mode: random
180
+ dtype: bfloat16
181
+ experts:
182
+ - source_model: v000000/L3.1-Storniitova-8B
183
+ - source_model: x0000001/l3.1-part_aaa
184
+
185
+ ```