froggeric commited on
Commit
bd8d0cb
1 Parent(s): 91c08e4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -2
README.md CHANGED
@@ -1,12 +1,58 @@
1
  ---
 
 
 
 
 
 
2
  license: apache-2.0
3
  language:
4
  - en
5
- library_name: transformers
6
  ---
 
 
 
 
 
 
 
 
 
 
 
 
7
 
 
8
 
9
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6585ffb10eeafbd678d4b3fe/jnqnl8a_zYYMqJoBpX8yS.png)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
 
11
  **Update Notes:**
12
  *Version 2 trained 1 additional epoch cycle for 3 total*
 
1
  ---
2
+ base_model:
3
+ - senseable/WestLake-7B-v2
4
+ library_name: transformers
5
+ tags:
6
+ - mergekit
7
+ - merge
8
  license: apache-2.0
9
  language:
10
  - en
 
11
  ---
12
+ # WestLake-10.7B-v2
13
+
14
+ This is my first viable self-merge of this fantastic model obtained after 12 rounds of testing with different
15
+ merge settings. In my benchmark, it greatly improves over the original 7B model, and ranks between miqu-1-120b
16
+ and goliath-120b! I would describe the improvements as a better writing style, with more details.
17
+
18
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
19
+
20
+ ## Merge Details
21
+ ### Merge Method
22
+
23
+ This model was merged using the passthrough merge method.
24
 
25
+ ### Models Merged
26
 
27
+ The following models were included in the merge:
28
+ * [senseable/WestLake-7B-v2](https://huggingface.co/senseable/WestLake-7B-v2)
29
+
30
+ ### Configuration
31
+
32
+ The following YAML configuration was used to produce this model:
33
+
34
+ ```yaml
35
+ dtype: float16
36
+ merge_method: passthrough
37
+ slices:
38
+ - sources:
39
+ - model: senseable/WestLake-7B-v2
40
+ layer_range: [0,9]
41
+ - sources:
42
+ - model: senseable/WestLake-7B-v2
43
+ layer_range: [5,14]
44
+ - sources:
45
+ - model: senseable/WestLake-7B-v2
46
+ layer_range: [10,19]
47
+ - sources:
48
+ - model: senseable/WestLake-7B-v2
49
+ layer_range: [15,24]
50
+ - sources:
51
+ - model: senseable/WestLake-7B-v2
52
+ layer_range: [20,32]
53
+ ```
54
+
55
+ ---
56
 
57
  **Update Notes:**
58
  *Version 2 trained 1 additional epoch cycle for 3 total*