add readme
Browse files
README.md
CHANGED
@@ -2,10 +2,37 @@
|
|
2 |
---
|
3 |
license: apache-2.0
|
4 |
---
|
|
|
|
|
5 |
# Seraph-7B
|
6 |
|
7 |
This is the model for Seraph-7B. I used [mergekit](https://github.com/cg123/mergekit) to merge models.
|
8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
# Yaml Config
|
10 |
|
11 |
```yaml
|
@@ -25,4 +52,18 @@ parameters:
|
|
25 |
value: [1, 0.5, 0.7, 0.3, 0]
|
26 |
- value: 0.5 # fallback for rest of tensors
|
27 |
dtype: bfloat16
|
28 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
---
|
3 |
license: apache-2.0
|
4 |
---
|
5 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/NjbXLTcpychqNC0Yt-1w-.png)
|
6 |
+
|
7 |
# Seraph-7B
|
8 |
|
9 |
This is the model for Seraph-7B. I used [mergekit](https://github.com/cg123/mergekit) to merge models.
|
10 |
|
11 |
+
# Prompt Templates
|
12 |
+
|
13 |
+
You can use these prompt templates, but I recommend using ChatML.
|
14 |
+
|
15 |
+
### ChatML:
|
16 |
+
|
17 |
+
```
|
18 |
+
<|im_start|>system
|
19 |
+
{system}<|im_end|>
|
20 |
+
<|im_start|>user
|
21 |
+
{user}<|im_end|>
|
22 |
+
<|im_start|>assistant
|
23 |
+
{asistant}<|im_end|>
|
24 |
+
```
|
25 |
+
|
26 |
+
### System, User, Asistant Alpaca Style:
|
27 |
+
|
28 |
+
```
|
29 |
+
### System:
|
30 |
+
{system}
|
31 |
+
### User:
|
32 |
+
{user}
|
33 |
+
### Assistant:
|
34 |
+
```
|
35 |
+
|
36 |
# Yaml Config
|
37 |
|
38 |
```yaml
|
|
|
52 |
value: [1, 0.5, 0.7, 0.3, 0]
|
53 |
- value: 0.5 # fallback for rest of tensors
|
54 |
dtype: bfloat16
|
55 |
+
```
|
56 |
+
|
57 |
+
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
58 |
+
|
59 |
+
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Seraph-7B)
|
60 |
+
|
61 |
+
| Metric | Value |
|
62 |
+
|-----------------------|---------------------------|
|
63 |
+
| Avg. | 71.86 |
|
64 |
+
| ARC (25-shot) | 67.83 |
|
65 |
+
| HellaSwag (10-shot) | 86.22 |
|
66 |
+
| MMLU (5-shot) | 65.07|
|
67 |
+
| TruthfulQA (0-shot) | 59.49 |
|
68 |
+
| Winogrande (5-shot) | 80.66 |
|
69 |
+
| GSM8K (5-shot) | 71.87 |
|