aashish1904 commited on
Commit
3eac2f8
·
verified ·
1 Parent(s): 68e2944

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +551 -0
README.md ADDED
@@ -0,0 +1,551 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ pipeline_tag: text-generation
5
+ library_name: transformers
6
+ language:
7
+ - en
8
+ license: llama3
9
+ tags:
10
+ - mergekit
11
+ - merge
12
+ - multi-step merge
13
+ - not-for-all-audiences
14
+ - nsfw
15
+ - rp
16
+ - roleplay
17
+ - role-play
18
+ - summarization
19
+ - emotion classification
20
+ base_model:
21
+ - nothingiisreal/L3-8B-Celeste-v1
22
+ - Nitral-AI/Hathor_Tahsin-L3-8B-v0.85
23
+ - Sao10K/L3-8B-Stheno-v3.2
24
+ - ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B
25
+ - Sao10K/L3-8B-Lunaris-v1
26
+ - turboderp/llama3-turbcat-instruct-8b
27
+ - ChaoticNeutrals/Domain-Fusion-L3-8B
28
+ - migtissera/Llama-3-8B-Synthia-v3.5
29
+ - TheDrummer/Llama-3SOME-8B-v2
30
+ - ChaoticNeutrals/Hathor_RP-v.01-L3-8B
31
+ - TheSkullery/llama-3-cat-8b-instruct-v1
32
+ - FPHam/L3-8B-Everything-COT
33
+ - Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged
34
+ - OEvortex/Emotional-llama-8B
35
+ - lighteternal/Llama3-merge-biomed-8b
36
+ - Casual-Autopsy/Llama3-merge-psychotherapy-8b
37
+ - Sao10K/L3-8B-Tamamo-v1
38
+ - ResplendentAI/Nymph_8B
39
+ - ChaoticNeutrals/T-900-8B
40
+ - Sao10K/L3-8B-Niitama-v1
41
+ - bluuwhale/L3-SthenoMaidBlackroot-8B-V1
42
+ - Hastagaras/Jamet-8B-L3-MK.V-Blackroot
43
+ - Hastagaras/Halu-8B-Llama3-Blackroot
44
+ - crestf411/L3-8B-sunfall-v0.4-stheno-v3.2
45
+
46
+ ---
47
+
48
+ ![](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)
49
+
50
+ # QuantFactory/L3-Super-Nova-RP-8B-GGUF
51
+ This is quantized version of [Casual-Autopsy/L3-Super-Nova-RP-8B](https://huggingface.co/Casual-Autopsy/L3-Super-Nova-RP-8B) created using llama.cpp
52
+
53
+ # Original Model Card
54
+
55
+ | <img src="https://huggingface.co/Casual-Autopsy/L3-Super-Nova-RP-8B/resolve/main/Card-Assets/NovaKid-Girl.jpeg" width="50%" height="50%" style="display: block; margin: auto;"> |
56
+ |:---:|
57
+ | Image generated by [mayonays_on_toast](https://civitai.com/user/mayonays_on_toast) - [Sauce](https://civitai.com/images/10153472) |
58
+ ***
59
+ ***
60
+ ***
61
+ # L3-Super-Nova-RP-8B
62
+
63
+ This is a role-playing model designed with the goal of good creativity and intelligence to improve advance role-playing experiences. The aim of L3-Super-Nova-RP-8B is to be good at Chain-of-Thoughts, summarizing information, and recognizing emotions. It also includes data about the human body and mind in an attempt to enhance understanding and interaction within role-playing scenarios.
64
+
65
+ The model was developed using various methods in multiple merging steps. To boost creativity, it used techniques to strengthen and adjust its output which was paried with the newly released merge method. All merge calculations were done in float32 format and then converted to the usual bfloat16 during merging.
66
+
67
+ ***
68
+ ***
69
+ ## Presets
70
+
71
+ ***
72
+ ### Text Gen
73
+ The current good starting preset for this model. **Subject to change.**
74
+ **Settings by yours truly**
75
+ ```yaml
76
+ Top K: 40
77
+ Min P: 0.075 # I've got some good results as low as 0.05 as well
78
+ Repetition Penalty: 1.01
79
+ # Don't make this higher, DRY handles the bulk of Squashing Repetition.
80
+ # This is just to lightly nudge the bot to move the plot forward
81
+ Rep Pen Range: 2048 # Don't make this higher either.
82
+ Presence Penalty: 0.03 # Minor encouragement to use synonyms. Don't make this higher maybe?
83
+ Smoothing Factor: 0.3
84
+
85
+ DRY Repetition Penalty:
86
+ Multiplier: 0.8
87
+ Base: 1.75
88
+ Allowed Length: 2
89
+ Penalty Range: 4096
90
+
91
+ Dynamic Temperature:
92
+ Min Temp: 0.5
93
+ Max Temp: 1.25
94
+ Exponent: 0.85
95
+ ```
96
+
97
+ ***
98
+ ### Context/Instruct
99
+ [Virt-io's SillyTavern Presets](https://huggingface.co/Virt-io/SillyTavern-Presets) work really well with this.
100
+
101
+ ***
102
+ ***
103
+ ## Usage Info
104
+
105
+ Some of the **INT** models were chosen with some of SillyTavern's features in mind, such as emotion based sprites, dynamic music, and pretty much any feature, extension, or STscript that uses sumarization. With that said, it's recommended to use SillyTavern as your front-end.
106
+
107
+ While not required, I'd recommend building the story string prompt with Lorebooks rather than using the Advance Formatting menu. The only thing you really need in the Story String prompt within Advance Formatting is the system prompt. Doing it this way tends to keep the character more consistent as the RP goes on as all character card info is locked to a certain depth rather than getting further and further away within the context.
108
+
109
+ ***
110
+ ***
111
+ ## Quants
112
+
113
+ GGUF:
114
+ - [Static GGUFs](https://huggingface.co/mradermacher/L3-Super-Nova-RP-8B-GGUF) by mradermacher
115
+ - [Imatrix GGUFs](https://huggingface.co/mradermacher/L3-Super-Nova-RP-8B-i1-GGUF) by mradermacher
116
+
117
+ Exl2:
118
+ - [8.0bpw-h8 Exl2](https://huggingface.co/Slvcxc/L3-Super-Nova-RP-8B-8.0bpw-h8-exl2) by Slvcxc
119
+
120
+ ***
121
+ ***
122
+ ## Merge Info
123
+
124
+ The merge methods used were **Ties**, **Dare Ties**, **Breadcrumbs Ties**, **SLERP**, and **DELLA**.
125
+
126
+ The model was finished off with both **Merge Densification**, and **Negative Weighting** techniques to boost creativity.
127
+
128
+ All merging steps had the merge calculations done in **float32** and were output as **bfloat16**.
129
+
130
+ ***
131
+ ### Models Merged
132
+
133
+ The following models were used to make this merge:
134
+ * [nothingiisreal/L3-8B-Celeste-v1](https://huggingface.co/nothingiisreal/L3-8B-Celeste-v1)
135
+ * [Nitral-AI/Hathor_Tahsin-L3-8B-v0.85](https://huggingface.co/Nitral-AI/Hathor_Tahsin-L3-8B-v0.85)
136
+ * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2)
137
+ * [ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B](https://huggingface.co/ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B)
138
+ * [Sao10K/L3-8B-Lunaris-v1](https://huggingface.co/Sao10K/L3-8B-Lunaris-v1)
139
+ * [turboderp/llama3-turbcat-instruct-8b](https://huggingface.co/turboderp/llama3-turbcat-instruct-8b)
140
+ * [ChaoticNeutrals/Domain-Fusion-L3-8B](https://huggingface.co/ChaoticNeutrals/Domain-Fusion-L3-8B)
141
+ * [migtissera/Llama-3-8B-Synthia-v3.5](https://huggingface.co/migtissera/Llama-3-8B-Synthia-v3.5)
142
+ * [TheDrummer/Llama-3SOME-8B-v2](https://huggingface.co/TheDrummer/Llama-3SOME-8B-v2)
143
+ * [ChaoticNeutrals/Hathor_RP-v.01-L3-8B](https://huggingface.co/ChaoticNeutrals/Hathor_RP-v.01-L3-8B)
144
+ * [TheSkullery/llama-3-cat-8b-instruct-v1](https://huggingface.co/TheSkullery/llama-3-cat-8b-instruct-v1)
145
+ * [FPHam/L3-8B-Everything-COT](https://huggingface.co/FPHam/L3-8B-Everything-COT)
146
+ * [Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged](https://huggingface.co/Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged)
147
+ * [OEvortex/Emotional-llama-8B](https://huggingface.co/OEvortex/Emotional-llama-8B)
148
+ * [lighteternal/Llama3-merge-biomed-8b](https://huggingface.co/lighteternal/Llama3-merge-biomed-8b)
149
+ * [Casual-Autopsy/Llama3-merge-psychotherapy-8b](https://huggingface.co/Casual-Autopsy/Llama3-merge-psychotherapy-8b)
150
+ * [Sao10K/L3-8B-Tamamo-v1](https://huggingface.co/Sao10K/L3-8B-Tamamo-v1)
151
+ * [ResplendentAI/Nymph_8B](https://huggingface.co/ResplendentAI/Nymph_8B)
152
+ * [ChaoticNeutrals/T-900-8B](https://huggingface.co/ChaoticNeutrals/T-900-8B)
153
+ * [Sao10K/L3-8B-Niitama-v1](https://huggingface.co/Sao10K/L3-8B-Niitama-v1)
154
+ * [bluuwhale/L3-SthenoMaidBlackroot-8B-V1](https://huggingface.co/bluuwhale/L3-SthenoMaidBlackroot-8B-V1)
155
+ * [Hastagaras/Jamet-8B-L3-MK.V-Blackroot](https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot)
156
+ * [Hastagaras/Halu-8B-Llama3-Blackroot](https://huggingface.co/Hastagaras/Halu-8B-Llama3-Blackroot)
157
+ * [crestf411/L3-8B-sunfall-v0.4-stheno-v3.2](https://huggingface.co/crestf411/L3-8B-sunfall-v0.4-stheno-v3.2)
158
+
159
+ ***
160
+ ***
161
+ ## Evaluation Results
162
+
163
+ ***
164
+ ### [Open LLM Leaderboard](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
165
+
166
+ **Explaination for AI RP newbies:** IFEval is the most important evaluation for RP AIs as it determines how well it can follow OOC, Lorebooks, and most importantly character cards.
167
+ The rest don't matter. At least not nearly as much as IFEval.
168
+
169
+ |Metric | Value|
170
+ |:------------------|------:|
171
+ |Avg. |N/A|
172
+ |IFEval (0-Shot) |N/A|
173
+ |BBH (3-Shot) |N/A|
174
+ |MATH Lvl 5 (4-Shot)|N/A|
175
+ |GPQA (0-shot) |N/A|
176
+ |MuSR (0-shot) |N/A|
177
+ |MMLU-PRO (5-shot) |N/A|
178
+
179
+ ***
180
+ ### [UGI Leaderboard](https://huggingface.co/spaces/DontPlanToEnd/UGI-Leaderboard)
181
+
182
+ Information about the metrics can be found at the bottom of the [UGI Leaderboard](https://huggingface.co/spaces/DontPlanToEnd/UGI-Leaderboard) in the respective tabs.
183
+
184
+ |Metric(UGI-Leaderboard) | Value | Value | Metric(Writing Style)|
185
+ |:------------------------|:-----:|:-----:|----------------------:|
186
+ |UGI(Avg.) |23.56 |0.199 |RegV1 |
187
+ |W/10 |5.8 |0.218 |RegV2 |
188
+ |Unruly |22.5 |0.15 |MyScore |
189
+ |Internet |11.8 |8.34 |ASSS |
190
+ |Stats |18.7 |10.26 |SMOG |
191
+ |Writing |31.5 |1.76 |Yule |
192
+ |PolContro |33.3 | | |
193
+
194
+ ***
195
+ ***
196
+ ## Secret Sauce
197
+
198
+ The following YAML configs were used to make this merge.
199
+
200
+ ***
201
+ ### Super-Nova-CRE_pt.1
202
+
203
+ ```yaml
204
+ models:
205
+ - model: nothingiisreal/L3-8B-Celeste-v1
206
+ - model: Nitral-AI/Hathor_Tahsin-L3-8B-v0.85
207
+ parameters:
208
+ density: [0.35, 0.45, 0.5, 0.55, 0.65, 0.55, 0.5, 0.45, 0.35]
209
+ weight: [0.495, 0.165, 0.165, 0.495, 0.495, 0.165, 0.165, 0.495]
210
+ - model: Sao10K/L3-8B-Stheno-v3.2
211
+ parameters:
212
+ density: [0.65, 0.55, 0.5, 0.45, 0.35, 0.45, 0.5, 0.55, 0.65]
213
+ weight: [0.165, 0.495, 0.495, 0.165, 0.165, 0.495, 0.495, 0.165]
214
+ merge_method: dare_ties
215
+ base_model: nothingiisreal/L3-8B-Celeste-v1
216
+ parameters:
217
+ normalize: false
218
+ int8_mask: true
219
+ dtype: float32
220
+ out_dtype: bfloat16
221
+ ```
222
+
223
+ ***
224
+ ### Super-Nova-CRE_pt.2
225
+
226
+ ```yaml
227
+ models:
228
+ - model: nothingiisreal/L3-8B-Celeste-v1
229
+ - model: ChaoticNeutrals/Poppy_Porpoise-1.0-L3-8B
230
+ parameters:
231
+ density: [0.35, 0.45, 0.5, 0.55, 0.65, 0.55, 0.5, 0.45, 0.35]
232
+ weight: [0.165, 0.495, 0.495, 0.165, 0.165, 0.495, 0.495, 0.165]
233
+ - model: Sao10K/L3-8B-Lunaris-v1
234
+ parameters:
235
+ density: [0.65, 0.55, 0.5, 0.45, 0.35, 0.45, 0.5, 0.55, 0.65]
236
+ weight: [0.495, 0.165, 0.165, 0.495, 0.495, 0.165, 0.165, 0.495]
237
+ merge_method: dare_ties
238
+ base_model: nothingiisreal/L3-8B-Celeste-v1
239
+ parameters:
240
+ normalize: false
241
+ int8_mask: true
242
+ dtype: float32
243
+ out_dtype: bfloat16
244
+ ```
245
+
246
+ ***
247
+ ### Super-Nova-UNC_pt.1
248
+
249
+ ```yaml
250
+ models:
251
+ - model: turboderp/llama3-turbcat-instruct-8b
252
+ - model: ChaoticNeutrals/Domain-Fusion-L3-8B
253
+ parameters:
254
+ density: 0.5
255
+ weight: [0.495, 0.165, 0.165, 0.495, 0.495, 0.165, 0.165, 0.495]
256
+ - model: migtissera/Llama-3-8B-Synthia-v3.5
257
+ parameters:
258
+ density: 0.5
259
+ weight: [0.165, 0.495, 0.495, 0.165, 0.165, 0.495, 0.495, 0.165]
260
+ merge_method: dare_ties
261
+ base_model: turboderp/llama3-turbcat-instruct-8b
262
+ parameters:
263
+ normalize: false
264
+ int8_mask: true
265
+ dtype: float32
266
+ out_dtype: bfloat16
267
+ ```
268
+
269
+ ***
270
+ ### Super-Nova-UNC_pt.2
271
+
272
+ ```yaml
273
+ models:
274
+ - model: turboderp/llama3-turbcat-instruct-8b
275
+ - model: TheDrummer/Llama-3SOME-8B-v2
276
+ parameters:
277
+ density: 0.5
278
+ weight: [0.165, 0.495, 0.495, 0.165, 0.165, 0.495, 0.495, 0.165]
279
+ - model: ChaoticNeutrals/Hathor_RP-v.01-L3-8B
280
+ parameters:
281
+ density: 0.5
282
+ weight: [0.495, 0.165, 0.165, 0.495, 0.495, 0.165, 0.165, 0.495]
283
+ merge_method: dare_ties
284
+ base_model: turboderp/llama3-turbcat-instruct-8b
285
+ parameters:
286
+ normalize: false
287
+ int8_mask: true
288
+ dtype: float32
289
+ out_dtype: bfloat16
290
+ ```
291
+
292
+ ***
293
+ ### Super-Nova-INT_pt.1
294
+
295
+ ```yaml
296
+ models:
297
+ - model: TheSkullery/llama-3-cat-8b-instruct-v1
298
+ - model: FPHam/L3-8B-Everything-COT
299
+ parameters:
300
+ density: 0.5
301
+ weight: [0.139, 0.139, 0.208, 0.139, 0.208]
302
+ - model: Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged
303
+ parameters:
304
+ density: 0.5
305
+ weight: [0.139, 0.208, 0.139, 0.208, 0.139]
306
+ - model: OEvortex/Emotional-llama-8B
307
+ parameters:
308
+ density: 0.5
309
+ weight: [0.208, 0.139, 0.208, 0.139, 0.139]
310
+ - model: lighteternal/Llama3-merge-biomed-8b
311
+ parameters:
312
+ density: 0.5
313
+ weight: [0.208, 0.139, 0.139, 0.139, 0.208]
314
+ - model: Casual-Autopsy/Llama3-merge-psychotherapy-8b
315
+ parameters:
316
+ density: 0.5
317
+ weight: [0.139, 0.208, 0.139, 0.208, 0.139]
318
+ merge_method: ties
319
+ base_model: TheSkullery/llama-3-cat-8b-instruct-v1
320
+ parameters:
321
+ normalize: false
322
+ int8_mask: true
323
+ dtype: float32
324
+ out_dtype: bfloat16
325
+ ```
326
+
327
+ ***
328
+ ### Super-Nova-INT_pt.2
329
+
330
+ ```yaml
331
+ models:
332
+ - model: TheSkullery/llama-3-cat-8b-instruct-v1
333
+ - model: FPHam/L3-8B-Everything-COT
334
+ parameters:
335
+ density: 0.9
336
+ gamma: 0.01
337
+ weight: [0.139, 0.208, 0.208, 0.139, 0.139]
338
+ - model: Ayush-1722/Meta-Llama-3-8B-Instruct-Summarize-v0.2-24K-LoRANET-Merged
339
+ parameters:
340
+ density: 0.9
341
+ gamma: 0.01
342
+ weight: [0.208, 0.139, 0.139, 0.139, 0.208]
343
+ - model: OEvortex/Emotional-llama-8B
344
+ parameters:
345
+ density: 0.9
346
+ gamma: 0.01
347
+ weight: [0.139, 0.139, 0.208, 0.208, 0.139]
348
+ - model: lighteternal/Llama3-merge-biomed-8b
349
+ parameters:
350
+ density: 0.9
351
+ gamma: 0.01
352
+ weight: [0.139, 0.208, 0.139, 0.208, 0.139]
353
+ - model: Casual-Autopsy/Llama3-merge-psychotherapy-8b
354
+ parameters:
355
+ density: 0.9
356
+ gamma: 0.01
357
+ weight: [0.208, 0.139, 0.139, 0.139, 0.208]
358
+ merge_method: breadcrumbs_ties
359
+ base_model: TheSkullery/llama-3-cat-8b-instruct-v1
360
+ parameters:
361
+ normalize: false
362
+ int8_mask: true
363
+ dtype: float32
364
+ out_dtype: bfloat16
365
+ ```
366
+
367
+ ***
368
+ ### Super-Nova-CRE
369
+
370
+ ```yaml
371
+
372
+ models:
373
+ - model: Casual-Autopsy/Super-Nova-CRE_pt.1
374
+ - model: Casual-Autopsy/Super-Nova-CRE_pt.2
375
+ merge_method: slerp
376
+ base_model: Casual-Autopsy/Super-Nova-CRE_pt.1
377
+ parameters:
378
+ t:
379
+ - filter: self_attn
380
+ value: [0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5]
381
+ - filter: mlp
382
+ value: [0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5]
383
+ - value: 0.5
384
+ embed_slerp: true
385
+ dtype: float32
386
+ out_dtype: bfloat16
387
+ ```
388
+
389
+ ***
390
+ ### Super-Nova-UNC
391
+
392
+ ```yaml
393
+ models:
394
+ - model: Casual-Autopsy/Super-Nova-UNC_pt.1
395
+ - model: Casual-Autopsy/Super-Nova-UNC_pt.2
396
+ merge_method: slerp
397
+ base_model: Casual-Autopsy/Super-Nova-UNC_pt.1
398
+ parameters:
399
+ t:
400
+ - value: [0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5, 0.7, 0.3, 0.5, 0.3, 0.7, 0.5]
401
+ embed_slerp: true
402
+ dtype: float32
403
+ out_dtype: bfloat16
404
+ ```
405
+
406
+ ***
407
+ ### Super-Nova-INT
408
+
409
+ ```yaml
410
+ models:
411
+ - model: Casual-Autopsy/Super-Nova-INT_pt.1
412
+ - model: Casual-Autopsy/Super-Nova-INT_pt.2
413
+ merge_method: slerp
414
+ base_model: Casual-Autopsy/Super-Nova-INT_pt.1
415
+ parameters:
416
+ t:
417
+ - value: 0.5
418
+ embed_slerp: true
419
+ dtype: float32
420
+ out_dtype: bfloat16
421
+ ```
422
+
423
+ ***
424
+ ### Super-Nova-RP_stp.1
425
+
426
+ ```yaml
427
+
428
+ models:
429
+ - model: Casual-Autopsy/Super-Nova-CRE
430
+ - model: asual-Autopsy/Super-Nova-UNC
431
+ merge_method: slerp
432
+ base_model: Casual-Autopsy/Super-Nova-CRE
433
+ parameters:
434
+ t:
435
+ - value: [0.7, 0.5, 0.3, 0.25, 0.2, 0.25, 0.3, 0.5, 0.7]
436
+ embed_slerp: true
437
+ dtype: float32
438
+ out_dtype: bfloat16
439
+
440
+ ```
441
+
442
+ ***
443
+ ### Super-Nova-RP_stp.2
444
+
445
+ ```yaml
446
+ models:
447
+ - model: Casual-Autopsy/Super-Nova-RP_stp.1
448
+ - model: Casual-Autopsy/Super-Nova-INT
449
+ merge_method: slerp
450
+ base_model: Casual-Autopsy/Super-Nova-RP_stp.1
451
+ parameters:
452
+ t:
453
+ - value: [0.1, 0.15, 0.2, 0.4, 0.6, 0.4, 0.2, 0.15, 0.1]
454
+ embed_slerp: true
455
+ dtype: float32
456
+ out_dtype: bfloat16
457
+ ```
458
+
459
+ ***
460
+ ### Super-Nova-RP_pt.1
461
+
462
+ ```yaml
463
+ models:
464
+ - model: Casual-Autopsy/Super-Nova-RP_stp.2
465
+ - model: Sao10K/L3-8B-Tamamo-v1
466
+ parameters:
467
+ density: [0.4, 0.6, 0.5, 0.6, 0.4]
468
+ epsilon: [0.15, 0.15, 0.25, 0.15, 0.15]
469
+ lambda: 0.85
470
+ weight: [-0.01523, 0.01768, -0.01384, 0.01835, -0.01247]
471
+ - model: ResplendentAI/Nymph_8B
472
+ parameters:
473
+ density: [0.65, 0.35, 0.5, 0.35, 0.65]
474
+ epsilon: [0.1, 0.1, 0.25, 0.1, 0.1]
475
+ lambda: 0.85
476
+ weight: [0.01823, -0.01647, 0.01422, -0.01975, 0.01128]
477
+ - model: ChaoticNeutrals/T-900-8B
478
+ parameters:
479
+ density: [0.35, 0.65, 0.5, 0.65, 0.35]
480
+ epsilon: [0.1, 0.1, 0.25, 0.1, 0.1]
481
+ lambda: 0.85
482
+ weight: [-0.01891, 0.01554, -0.01325, 0.01791, -0.01458]
483
+ - model: Sao10K/L3-8B-Niitama-v1
484
+ parameters:
485
+ density: [0.6, 0.4, 0.5, 0.4, 0.6]
486
+ epsilon: [0.15, 0.15, 0.25, 0.15, 0.15]
487
+ lambda: 0.85
488
+ weight: [0.01768, -0.01675, 0.01285, -0.01696, 0.01421]
489
+ merge_method: della
490
+ base_model: Casual-Autopsy/Super-Nova-RP_stp.2
491
+ parameters:
492
+ normalize: false
493
+ int8_mask: true
494
+ dtype: float32
495
+ out_dtype: bfloat16
496
+ ```
497
+
498
+ ***
499
+ ### Super-Nova-RP_pt.2
500
+
501
+ ```yaml
502
+ models:
503
+ - model: Casual-Autopsy/Super-Nova-RP_stp.2
504
+ - model: bluuwhale/L3-SthenoMaidBlackroot-8B-V1
505
+ parameters:
506
+ density: [0.4, 0.6, 0.5, 0.6, 0.4]
507
+ epsilon: [0.15, 0.15, 0.25, 0.15, 0.15]
508
+ lambda: 0.85
509
+ weight: [-0.01935, 0.01785, -0.01512, 0.01809, -0.01371]
510
+ - model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
511
+ parameters:
512
+ density: [0.65, 0.35, 0.5, 0.35, 0.65]
513
+ epsilon: [0.1, 0.1, 0.25, 0.1, 0.1]
514
+ lambda: 0.85
515
+ weight: [0.01847, -0.01468, 0.01503, -0.01822, 0.01459]
516
+ - model: Hastagaras/Halu-8B-Llama3-Blackroot
517
+ parameters:
518
+ density: [0.35, 0.65, 0.5, 0.65, 0.35]
519
+ epsilon: [0.1, 0.1, 0.25, 0.1, 0.1]
520
+ lambda: 0.85
521
+ weight: [-0.01578, 0.01821, -0.01753, 0.01677, -0.01442]
522
+ - model: crestf411/L3-8B-sunfall-v0.4-stheno-v3.2
523
+ parameters:
524
+ density: [0.6, 0.5, 0.5, 0.5, 0.6]
525
+ epsilon: [0.15, 0.15, 0.25, 0.15, 0.15]
526
+ lambda: 0.85
527
+ weight: [0.01667, -0.01740, 0.01560, -0.01564, 0.01315]
528
+ merge_method: della
529
+ base_model: Casual-Autopsy/Super-Nova-RP_stp.2
530
+ parameters:
531
+ normalize: false
532
+ int8_mask: true
533
+ dtype: float32
534
+ out_dtype: bfloat16
535
+ ```
536
+
537
+ ***
538
+ ### L3-Super-Nova-RP-8B
539
+
540
+ ```yaml
541
+ models:
542
+ - model: Casual-Autopsy/Super-Nova-RP_pt.1
543
+ - model: Casual-Autopsy/Super-Nova-RP_pt.2
544
+ merge_method: slerp
545
+ base_model: Casual-Autopsy/Super-Nova-RP_pt.1
546
+ parameters:
547
+ t:
548
+ - value: 0.5
549
+ dtype: float32
550
+ out_dtype: bfloat16
551
+ ```