saishf commited on
Commit
1955a2b
1 Parent(s): f393f33

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +91 -0
README.md ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - ammarali32/multi_verse_model
4
+ - jeiku/Theory_of_Mind_Roleplay_Mistral
5
+ - ammarali32/multi_verse_model
6
+ - jeiku/Alpaca_NSFW_Shuffled_Mistral
7
+ - ammarali32/multi_verse_model
8
+ - jeiku/Theory_of_Mind_Mistral
9
+ - ammarali32/multi_verse_model
10
+ - jeiku/Gnosis_Reformatted_Mistral
11
+ - ammarali32/multi_verse_model
12
+ - ammarali32/multi_verse_model
13
+ - jeiku/Re-Host_Limarp_Mistral
14
+ - ammarali32/multi_verse_model
15
+ - jeiku/Luna_LoRA_Mistral
16
+ library_name: transformers
17
+ license: cc-by-nc-4.0
18
+ tags:
19
+ - mergekit
20
+ - merge
21
+ language:
22
+ - en
23
+ ---
24
+
25
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/63df7c44f0c75dfb876272c0/EjMKmAvmKoKd_Z7o0b3VK.jpeg)
26
+ Multi verse img!
27
+ # merge
28
+
29
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
30
+
31
+ ## Merge Details
32
+ * This merge is entirely experimental, I've only tested it a few times but it seems to work? Thanks for all the loras jeiku. I keep getting driver crashes training my own :\
33
+ * Update, It scores well! My highest scoring model so far
34
+ ### Merge Method
35
+
36
+ This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [ammarali32/multi_verse_model](https://huggingface.co/ammarali32/multi_verse_model) as a base.
37
+
38
+ ### Models Merged
39
+
40
+ The following models were included in the merge:
41
+ * [ammarali32/multi_verse_model](https://huggingface.co/ammarali32/multi_verse_model) + [jeiku/Theory_of_Mind_Roleplay_Mistral](https://huggingface.co/jeiku/Theory_of_Mind_Roleplay_Mistral)
42
+ * [ammarali32/multi_verse_model](https://huggingface.co/ammarali32/multi_verse_model) + [jeiku/Alpaca_NSFW_Shuffled_Mistral](https://huggingface.co/jeiku/Alpaca_NSFW_Shuffled_Mistral)
43
+ * [ammarali32/multi_verse_model](https://huggingface.co/ammarali32/multi_verse_model) + [jeiku/Theory_of_Mind_Mistral](https://huggingface.co/jeiku/Theory_of_Mind_Mistral)
44
+ * [ammarali32/multi_verse_model](https://huggingface.co/ammarali32/multi_verse_model) + [jeiku/Gnosis_Reformatted_Mistral](https://huggingface.co/jeiku/Gnosis_Reformatted_Mistral)
45
+ * [ammarali32/multi_verse_model](https://huggingface.co/ammarali32/multi_verse_model) + [jeiku/Re-Host_Limarp_Mistral](https://huggingface.co/jeiku/Re-Host_Limarp_Mistral)
46
+ * [ammarali32/multi_verse_model](https://huggingface.co/ammarali32/multi_verse_model) + [jeiku/Luna_LoRA_Mistral](https://huggingface.co/jeiku/Luna_LoRA_Mistral)
47
+
48
+ ### Configuration
49
+
50
+ The following YAML configuration was used to produce this model:
51
+
52
+ ```yaml
53
+ merge_method: task_arithmetic
54
+ base_model: ammarali32/multi_verse_model
55
+ parameters:
56
+ normalize: true
57
+ models:
58
+ - model: ammarali32/multi_verse_model+jeiku/Gnosis_Reformatted_Mistral
59
+ parameters:
60
+ weight: 0.7
61
+ - model: ammarali32/multi_verse_model+jeiku/Theory_of_Mind_Roleplay_Mistral
62
+ parameters:
63
+ weight: 0.65
64
+ - model: ammarali32/multi_verse_model+jeiku/Luna_LoRA_Mistral
65
+ parameters:
66
+ weight: 0.5
67
+ - model: ammarali32/multi_verse_model+jeiku/Re-Host_Limarp_Mistral
68
+ parameters:
69
+ weight: 0.8
70
+ - model: ammarali32/multi_verse_model+jeiku/Alpaca_NSFW_Shuffled_Mistral
71
+ parameters:
72
+ weight: 0.75
73
+ - model: ammarali32/multi_verse_model+jeiku/Theory_of_Mind_Mistral
74
+ parameters:
75
+ weight: 0.7
76
+ dtype: float16
77
+
78
+ ```
79
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
80
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_saishf__Multi-Verse-RP-7B)
81
+
82
+ | Metric |Value|
83
+ |---------------------------------|----:|
84
+ |Avg. |74.73|
85
+ |AI2 Reasoning Challenge (25-Shot)|72.35|
86
+ |HellaSwag (10-Shot) |88.37|
87
+ |MMLU (5-Shot) |63.94|
88
+ |TruthfulQA (0-shot) |73.19|
89
+ |Winogrande (5-shot) |84.14|
90
+ |GSM8k (5-shot) |66.41|
91
+