aashish1904 commited on
Commit
132a4da
1 Parent(s): b939998

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +374 -0
README.md ADDED
@@ -0,0 +1,374 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ language:
5
+ - en
6
+ license: llama3
7
+ library_name: transformers
8
+ tags:
9
+ - merge
10
+ - mergekit
11
+ - lazymergekit
12
+ - not-for-all-audiences
13
+ - nsfw
14
+ - rp
15
+ - roleplay
16
+ - role-play
17
+ base_model:
18
+ - Casual-Autopsy/L3-Umbral-Mind-RP-v1.0-8B
19
+ - bluuwhale/L3-SthenoMaidBlackroot-8B-V1
20
+ - Cas-Warehouse/Llama-3-SOVL-MopeyMule-Blackroot-8B
21
+ - Cas-Warehouse/Llama-3-Mopeyfied-Psychology-v2
22
+ - tannedbum/L3-Nymeria-8B
23
+ - migtissera/Llama-3-8B-Synthia-v3.5
24
+ - Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
25
+ - tannedbum/L3-Nymeria-Maid-8B
26
+ - Casual-Autopsy/L3-Uncen-Merger-Omelette-RP-v0.2-8B
27
+ - aifeifei798/llama3-8B-DarkIdol-2.1-Uncensored-32K
28
+ - Nitral-AI/Hathor_Stable-v0.2-L3-8B
29
+ - Sao10K/L3-8B-Stheno-v3.1
30
+ pipeline_tag: text-generation
31
+ model-index:
32
+ - name: L3-Umbral-Mind-RP-v2.0-8B
33
+ results:
34
+ - task:
35
+ type: text-generation
36
+ name: Text Generation
37
+ dataset:
38
+ name: IFEval (0-Shot)
39
+ type: HuggingFaceH4/ifeval
40
+ args:
41
+ num_few_shot: 0
42
+ metrics:
43
+ - type: inst_level_strict_acc and prompt_level_strict_acc
44
+ value: 71.23
45
+ name: strict accuracy
46
+ source:
47
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
48
+ name: Open LLM Leaderboard
49
+ - task:
50
+ type: text-generation
51
+ name: Text Generation
52
+ dataset:
53
+ name: BBH (3-Shot)
54
+ type: BBH
55
+ args:
56
+ num_few_shot: 3
57
+ metrics:
58
+ - type: acc_norm
59
+ value: 32.49
60
+ name: normalized accuracy
61
+ source:
62
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
63
+ name: Open LLM Leaderboard
64
+ - task:
65
+ type: text-generation
66
+ name: Text Generation
67
+ dataset:
68
+ name: MATH Lvl 5 (4-Shot)
69
+ type: hendrycks/competition_math
70
+ args:
71
+ num_few_shot: 4
72
+ metrics:
73
+ - type: exact_match
74
+ value: 10.12
75
+ name: exact match
76
+ source:
77
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
78
+ name: Open LLM Leaderboard
79
+ - task:
80
+ type: text-generation
81
+ name: Text Generation
82
+ dataset:
83
+ name: GPQA (0-shot)
84
+ type: Idavidrein/gpqa
85
+ args:
86
+ num_few_shot: 0
87
+ metrics:
88
+ - type: acc_norm
89
+ value: 4.92
90
+ name: acc_norm
91
+ source:
92
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
93
+ name: Open LLM Leaderboard
94
+ - task:
95
+ type: text-generation
96
+ name: Text Generation
97
+ dataset:
98
+ name: MuSR (0-shot)
99
+ type: TAUR-Lab/MuSR
100
+ args:
101
+ num_few_shot: 0
102
+ metrics:
103
+ - type: acc_norm
104
+ value: 5.55
105
+ name: acc_norm
106
+ source:
107
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
108
+ name: Open LLM Leaderboard
109
+ - task:
110
+ type: text-generation
111
+ name: Text Generation
112
+ dataset:
113
+ name: MMLU-PRO (5-shot)
114
+ type: TIGER-Lab/MMLU-Pro
115
+ config: main
116
+ split: test
117
+ args:
118
+ num_few_shot: 5
119
+ metrics:
120
+ - type: acc
121
+ value: 30.26
122
+ name: accuracy
123
+ source:
124
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
125
+ name: Open LLM Leaderboard
126
+
127
+ ---
128
+
129
+ ![](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)
130
+
131
+ # QuantFactory/L3-Umbral-Mind-RP-v2.0-8B-GGUF
132
+ This is quantized version of [Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B](https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B) created using llama.cpp
133
+
134
+ # Original Model Card
135
+
136
+
137
+ | <img src="https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v3-8B/resolve/main/63073798_p0_master1200.jpg" style="display: block; margin: auto;"> |
138
+ |:---:|
139
+ | Image by ろ47 |
140
+ | |
141
+
142
+ # Merge
143
+
144
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
145
+
146
+ ***
147
+ ## Merge Details
148
+
149
+ The goal of this merge was to make an RP model better suited for role-plays with heavy themes such as but not limited to:
150
+ - Mental illness
151
+ - Self-harm
152
+ - Trauma
153
+ - Suicide
154
+
155
+ I hated how RP models tended to be overly positive and hopeful with role-plays involving such themes,
156
+ but thanks to [failspy/Llama-3-8B-Instruct-MopeyMule](https://huggingface.co/failspy/Llama-3-8B-Instruct-MopeyMule) this problem has been lessened considerably.
157
+
158
+ If you're an enjoyer of savior/reverse savior type role-plays like myself, then this model is for you.
159
+
160
+ ***
161
+ ## Usage Info
162
+
163
+ This model is meant to be used with asterisks/quotes RPing formats, any other format that isn't asterisks/quotes is likely to cause issues
164
+
165
+ ***
166
+ ## Quants
167
+
168
+ * [imatrix quants](https://huggingface.co/mradermacher/L3-Umbral-Mind-RP-v2.0-8B-i1-GGUF) by mradermacher
169
+ * [Static quants](https://huggingface.co/mradermacher/L3-Umbral-Mind-RP-v2.0-8B-GGUF) by mradermacher
170
+
171
+ * Exl2:
172
+ - [L3-Umbral-Mind-RP-v2.0-8B-8bpw-h8-exl2](https://huggingface.co/riveRiPH/L3-Umbral-Mind-RP-v2.0-8B-8bpw-h8-exl2) by riveRiPH
173
+ - [L3-Umbral-Mind-RP-v2.0-8B-6.3bpw-h8-exl2](https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B-6.3bpw-h8-exl2) by yours truly
174
+ - [L3-Umbral-Mind-RP-v2.0-8B-5.3bpw-h6-exl2](https://huggingface.co/riveRiPH/L3-Umbral-Mind-RP-v2.0-8B-5.3bpw-h6-exl2) by riveRiPH
175
+
176
+ ***
177
+ ## Merge Method
178
+
179
+ This model was merged using several Task Arithmetic merges and then tied together with a Model Stock merge, followed by another Task Arithmetic merge with a model containing psychology data.
180
+
181
+ ### Models Merged
182
+
183
+ The following models were included in the merge:
184
+ * [Casual-Autopsy/L3-Umbral-Mind-RP-v1.0-8B](https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v1.0-8B)
185
+ * [bluuwhale/L3-SthenoMaidBlackroot-8B-V1](https://huggingface.co/bluuwhale/L3-SthenoMaidBlackroot-8B-V1)
186
+ * [Cas-Warehouse/Llama-3-SOVL-MopeyMule-Blackroot-8B](https://huggingface.co/Cas-Warehouse/Llama-3-SOVL-MopeyMule-Blackroot-8B)
187
+ * [Cas-Warehouse/Llama-3-Mopeyfied-Psychology-v2](https://huggingface.co/Cas-Warehouse/Llama-3-Mopeyfied-Psychology-v2)
188
+ * [tannedbum/L3-Nymeria-8B](https://huggingface.co/tannedbum/L3-Nymeria-8B)
189
+ * [migtissera/Llama-3-8B-Synthia-v3.5](https://huggingface.co/migtissera/Llama-3-8B-Synthia-v3.5)
190
+ * [Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B](https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B)
191
+ * [tannedbum/L3-Nymeria-Maid-8B](https://huggingface.co/tannedbum/L3-Nymeria-Maid-8B)
192
+ * [Casual-Autopsy/L3-Uncen-Merger-Omelette-RP-v0.2-8B](https://huggingface.co/Casual-Autopsy/L3-Uncen-Merger-Omelette-RP-v0.2-8B)
193
+ * [aifeifei798/llama3-8B-DarkIdol-2.1-Uncensored-32K](https://huggingface.co/aifeifei798/llama3-8B-DarkIdol-2.1-Uncensored-32K)
194
+ * [Nitral-AI/Hathor_Stable-v0.2-L3-8B](https://huggingface.co/Nitral-AI/Hathor_Stable-v0.2-L3-8B)
195
+ * [Sao10K/L3-8B-Stheno-v3.1](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.1)
196
+
197
+ ***
198
+ ## Evaluation Results
199
+
200
+ ### [Open LLM Leaderboard](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
201
+
202
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Casual-Autopsy__L3-Umbral-Mind-RP-v2.0-8B)
203
+
204
+ **Explaination for AI RP newbies:** IFEval is the most important evaluation for RP AIs as it determines how well it can follow OOC, Lorebooks, and most importantly character cards.
205
+ The rest don't matter. At least not nearly as much as IFEval.
206
+
207
+ | Metric |Value|
208
+ |-------------------|----:|
209
+ |Avg. |25.76|
210
+ |IFEval (0-Shot) |71.23|
211
+ |BBH (3-Shot) |32.49|
212
+ |MATH Lvl 5 (4-Shot)|10.12|
213
+ |GPQA (0-shot) | 4.92|
214
+ |MuSR (0-shot) | 5.55|
215
+ |MMLU-PRO (5-shot) |30.26|
216
+
217
+ ### [UGI Leaderboard](https://huggingface.co/spaces/DontPlanToEnd/UGI-Leaderboard)
218
+
219
+ Information about the metrics can be found at the bottom of the [UGI Leaderboard](https://huggingface.co/spaces/DontPlanToEnd/UGI-Leaderboard) in the respective tabs.
220
+
221
+ |Metric(UGI-Leaderboard) | Value | Value | Metric(Writing Style)|
222
+ |:------------------------|:-----:|:-----:|----------------------:|
223
+ |UGI(Avg.) |31.82 |0.107 |RegV1 |
224
+ |W/10 |5.83 |0.096 |RegV2 |
225
+ |Unruly |43.3 |0.05 |MyScore |
226
+ |Internet |20 |9.12 |ASSS |
227
+ |Stats |23.6 |0 |SMOG |
228
+ |Writing |33.8 |1.47 |Yule |
229
+ |PolContro |38.3 | | |
230
+
231
+ ***
232
+ ## Secret Sauce
233
+
234
+ The following YAML configurations were used to produce this model:
235
+
236
+ ### Umbral-1
237
+
238
+ ```yaml
239
+ models:
240
+ - model: Casual-Autopsy/L3-Umbral-Mind-RP-v1.0-8B
241
+ - model: bluuwhale/L3-SthenoMaidBlackroot-8B-V1
242
+ parameters:
243
+ density: 0.45
244
+ weight: 0.4
245
+ - model: Cas-Warehouse/Llama-3-SOVL-MopeyMule-Blackroot-8B
246
+ parameters:
247
+ density: 0.65
248
+ weight: 0.1
249
+ merge_method: dare_ties
250
+ base_model: Casual-Autopsy/L3-Umbral-Mind-RP-v1.0-8B
251
+ parameters:
252
+ int8_mask: true
253
+ dtype: bfloat16
254
+ ```
255
+
256
+ ### Umbral-2
257
+
258
+ ```yaml
259
+ models:
260
+ - model: Cas-Warehouse/Llama-3-Mopeyfied-Psychology-v2
261
+ - model: tannedbum/L3-Nymeria-8B
262
+ parameters:
263
+ density: 0.45
264
+ weight: 0.25
265
+ - model: migtissera/Llama-3-8B-Synthia-v3.5
266
+ parameters:
267
+ density: 0.65
268
+ weight: 0.25
269
+ merge_method: dare_ties
270
+ base_model: Cas-Warehouse/Llama-3-Mopeyfied-Psychology-v2
271
+ parameters:
272
+ int8_mask: true
273
+ dtype: bfloat16
274
+ ```
275
+
276
+ ### Umbral-3
277
+
278
+ ```yaml
279
+ models:
280
+ - model: Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
281
+ - model: tannedbum/L3-Nymeria-Maid-8B
282
+ parameters:
283
+ density: 0.4
284
+ weight: 0.3
285
+ - model: Cas-Warehouse/Llama-3-SOVL-MopeyMule-Blackroot-8B
286
+ parameters:
287
+ density: 0.6
288
+ weight: 0.2
289
+ merge_method: dare_ties
290
+ base_model: Casual-Autopsy/L3-Umbral-Mind-RP-v0.3-8B
291
+ parameters:
292
+ int8_mask: true
293
+ dtype: bfloat16
294
+ ```
295
+
296
+ ### Mopey-Omelette
297
+
298
+ ```yaml
299
+ models:
300
+ - model: Casual-Autopsy/L3-Uncen-Merger-Omelette-RP-v0.2-8B
301
+ - model: Cas-Warehouse/Llama-3-SOVL-MopeyMule-Blackroot-8B
302
+ parameters:
303
+ weight: 0.15
304
+ merge_method: task_arithmetic
305
+ base_model: Casual-Autopsy/L3-Uncen-Merger-Omelette-RP-v0.2-8B
306
+ dtype: bfloat16
307
+ ```
308
+
309
+ ### Umbral-Mind-1
310
+
311
+ ```yaml
312
+ models:
313
+ - model: Casual-Autopsy/Umbral-1
314
+ - model: Casual-Autopsy/Umbral-3
315
+ merge_method: slerp
316
+ base_model: Casual-Autopsy/Umbral-1
317
+ parameters:
318
+ t:
319
+ - value: [0.7, 0.5, 0.3, 0.25, 0.2, 0.25, 0.3, 0.5, 0.7]
320
+ embed_slerp: true
321
+ dtype: bfloat16
322
+ ```
323
+
324
+ ### Umbral-Mind-2
325
+
326
+ ```yaml
327
+ models:
328
+ - model: Casual-Autopsy/Umbral-Mind-1
329
+ - model: Casual-Autopsy/Umbral-2
330
+ merge_method: slerp
331
+ base_model: Casual-Autopsy/Umbral-Mind-1
332
+ parameters:
333
+ t:
334
+ - value: [0.1, 0.15, 0.2, 0.4, 0.6, 0.4, 0.2, 0.15, 0.1]
335
+ embed_slerp: true
336
+ dtype: bfloat16
337
+ ```
338
+
339
+ ### Umbral-Mind-3
340
+
341
+ ```yaml
342
+ models:
343
+ - model: Casual-Autopsy/Umbral-Mind-2
344
+ - model: Casual-Autopsy/Mopey-Omelette
345
+ merge_method: slerp
346
+ base_model: Casual-Autopsy/Umbral-Mind-2
347
+ parameters:
348
+ t:
349
+ - value: [0.2, 0.25, 0.3, 0.4, 0.3, 0.25, 0.2, 0.25, 0.3, 0.4, 0.3, 0.25, 0.2]
350
+ embed_slerp: true
351
+ dtype: bfloat16
352
+ ```
353
+
354
+ ### L3-Umbral-Mind-RP-v2.0-8B
355
+
356
+ ```yaml
357
+ models:
358
+ - model: Casual-Autopsy/Umbral-Mind-3
359
+ - model: Casual-Autopsy/L3-Umbral-Mind-RP-v1.0-8B
360
+ parameters:
361
+ weight: 0.04
362
+ - model: aifeifei798/llama3-8B-DarkIdol-2.1-Uncensored-32K
363
+ parameters:
364
+ weight: 0.02
365
+ - model: Nitral-AI/Hathor_Stable-v0.2-L3-8B
366
+ parameters:
367
+ weight: 0.02
368
+ - model: Sao10K/L3-8B-Stheno-v3.1
369
+ parameters:
370
+ weight: 0.01
371
+ merge_method: task_arithmetic
372
+ base_model: Casual-Autopsy/Umbral-Mind-3
373
+ dtype: bfloat16
374
+ ```