SanjiWatsuki commited on
Commit
f2015a5
1 Parent(s): 3703cd7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +72 -0
README.md CHANGED
@@ -1,3 +1,75 @@
1
  ---
2
  license: cc-by-nc-4.0
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: cc-by-nc-4.0
3
+ tags:
4
+ - merge
5
  ---
6
+
7
+ ![image/png](https://huggingface.co/SanjiWatsuki/Loyal-Toppy-Bruins-Maid-7B-DARE/resolve/main/bruins-maid.png)
8
+
9
+ <!-- description start -->
10
+ ## Description
11
+
12
+ This repository hosts FP16 files for **Loyal-Toppy-Bruins-Maid-7B**, a 7B model aimed at having engaging RP with solid character card adherence and being a smart cookie at the same time.
13
+
14
+ Its foundation is [Starling-LM-7B-alpha](https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha), notable for its performance in the LMSYS Chatbot Arena, even surpassing GPT-3.5-Turbo-1106. The model incorporates [rwitz/go-bruins-v2](https://huggingface.co/rwitz/go-bruins-v2), a [Q-bert/MetaMath-Cybertron-Starling](https://huggingface.co/Q-bert/MetaMath-Cybertron-Starling) derivative with Alpaca RP data tuning.
15
+
16
+ The other foundational model is [chargoddard/loyal-piano-m7](https://huggingface.co/chargoddard/loyal-piano-m7), chosen for its strong RP performance and Alpaca format training, with a diverse dataset including PIPPA, rpbuild, and LimaRP.
17
+
18
+ [Undi95/Toppy-M-7B](https://huggingface.co/Undi95/Toppy-M-7B), known for its creativity, brings in useful RP data from various sources. It ranks first among 7B models on [OpenRouter](https://openrouter.ai/rankings) for a good reason.
19
+
20
+ [NeverSleep/Noromaid-7b-v0.1.1](https://huggingface.co/NeverSleep/Noromaid-7b-v0.1.1), a Mistral finetune with unique RP data not present in other models, was also added for bringing in a unique RP dataset and being a well-regarded RP model.
21
+
22
+ The models were merged using the DARE ties method, with a targeted 1.2 absolute weight and high density (0.5-0.6), as discussed in the [MergeKit GitHub Repo](https://github.com/cg123/mergekit/issues/26).
23
+
24
+ Currently, this model ranks at the top of my personal RP unit test benchmark and scored a very solid 20 on [lilblam's LLM Logic Test](https://docs.google.com/spreadsheets/d/1NgHDxbVWJFolq8bLvLkuPWKC7i_R6I6W/edit#gid=1278290632). My first impressions of it for RPing are very good but, admittedly, this model came out of the oven today so I haven't played it with it too much 😊
25
+
26
+ ### The sauce
27
+ ```
28
+ models: # Top-Loyal-Bruins-Maid-DARE-7B_v2
29
+ - model: mistralai/Mistral-7B-v0.1
30
+ # no parameters necessary for base model
31
+ - model: rwitz/go-bruins-v2 # MetamathCybertronStarling base
32
+ parameters:
33
+ weight: 0.5
34
+ density: 0.6
35
+ - model: chargoddard/loyal-piano-m7 # Pull in some PIPPA/LimaRP/Orca/rpguild
36
+ parameters:
37
+ weight: 0.5
38
+ density: 0.6
39
+ - model: Undi95/Toppy-M-7B
40
+ parameters:
41
+ weight: 0.1
42
+ density: 0.5
43
+ - model: NeverSleep/Noromaid-7b-v0.1.1
44
+ parameters:
45
+ weight: 0.1
46
+ density: 0.5
47
+ merge_method: dare_ties
48
+ base_model: mistralai/Mistral-7B-v0.1
49
+ parameters:
50
+ normalize: false
51
+ int8_mask: true
52
+ dtype: bfloat16
53
+ ```
54
+
55
+ <!-- description end -->
56
+ <!-- prompt-template start -->
57
+ ## Prompt template: Custom format, or Alpaca
58
+
59
+ ### Custom format:
60
+ I found the best SillyTavern results from using the Noromaid template.
61
+
62
+ SillyTavern config files: [Context](https://files.catbox.moe/ifmhai.json), [Instruct](https://files.catbox.moe/ttw1l9.json).
63
+
64
+ Otherwise, I tried to ensure that all of the underlying merged models were Alpaca favored.
65
+
66
+ ### Alpaca:
67
+ ```
68
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
69
+
70
+ ### Instruction:
71
+ {prompt}
72
+
73
+ ### Response:
74
+
75
+ ```