er1123090 commited on
Commit
e1a118b
1 Parent(s): 72e952b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +73 -0
README.md CHANGED
@@ -1,3 +1,76 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+
5
+ language:
6
+ - ko
7
+ base_model:
8
+ - LDCC/LDCC-SOLAR-10.7B
9
+ - hyeogi/SOLAR-10.7B-dpo-v1
10
+ tags:
11
+ - mergekit
12
+ - merge
13
+ - LDCC/LDCC-SOLAR-10.7B
14
+ - hyeogi/SOLAR-10.7B-dpo-v1
15
+ license: apache-2.0
16
+ ---
17
+ # merge
18
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
19
+
20
+ ## Merge Details
21
+ ### Merge Method
22
+
23
+ This model was merged using the SLERP merge method.
24
+
25
+ ### Models Merged
26
+
27
+ The following models were included in the merge:
28
+ * [hyeogi/SOLAR-10.7B-dpo-v1](https://huggingface.co/hyeogi/SOLAR-10.7B-dpo-v1)
29
+ * [LDCC/LDCC-SOLAR-10.7B](https://huggingface.co/LDCC/LDCC-SOLAR-10.7B)
30
+
31
+ ### Configuration
32
+
33
+ The following YAML configuration was used to produce this model:
34
+
35
+ ```yaml
36
+
37
+
38
+ slices:
39
+
40
+ - sources:
41
+
42
+ - model: LDCC/LDCC-SOLAR-10.7B
43
+
44
+ layer_range: [0, 48]
45
+
46
+ - model: hyeogi/SOLAR-10.7B-dpo-v1
47
+
48
+ layer_range: [0, 48]
49
+
50
+ merge_method: slerp
51
+ tokenizer_source: base
52
+ base_model: LDCC/LDCC-SOLAR-10.7B
53
+
54
+ embed_slerp: true
55
+
56
+ parameters:
57
+
58
+ t:
59
+
60
+ - filter: self_attn
61
+
62
+ value: [0, 0.5, 0.3, 0.7, 1]
63
+
64
+ - filter: mlp
65
+
66
+ value: [1, 0.5, 0.7, 0.3, 0]
67
+
68
+ - value: 0.5
69
+
70
+ dtype: bfloat16
71
+
72
+
73
+ ```
74
+ ## Datasets
75
+
76
+ Finetuned using LoRA with [kyujinpy/OpenOrca-KO](https://huggingface.co/datasets/kyujinpy/OpenOrca-KO)