kyujinpy commited on
Commit
2ba7fa0
1 Parent(s): 3a9e5f0

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +69 -0
README.md CHANGED
@@ -1,3 +1,72 @@
1
  ---
 
 
 
 
 
 
2
  license: cc-by-nc-sa-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ datasets:
5
+ - kyujinpy/Open-platypus-Commercial
6
+ library_name: transformers
7
+ pipeline_tag: text-generation
8
  license: cc-by-nc-sa-4.0
9
  ---
10
+
11
+ # **SOLAR-Platypus-10.7B-v1**
12
+
13
+ ## Model Details
14
+
15
+ **Model Developers** Kyujin Han (kyujinpy)
16
+
17
+ **Input** Models input text only.
18
+
19
+ **Output** Models generate text only.
20
+
21
+ **Model Architecture**
22
+ SOLAR-Platypus-10.7B-v1 is an auto-regressive language model based on the Llama2 architecture.
23
+
24
+ **Base Model**
25
+ [upstage/SOLAR-10.7B-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-v1.0)
26
+
27
+ **Training Dataset**
28
+ [kyujinpy/Open-platypus-Commercial](https://huggingface.co/datasets/kyujinpy/Open-platypus-Commercial).
29
+
30
+
31
+ ## Notice
32
+ While training, I used LoRA.
33
+ The lora_r values is 16.
34
+
35
+ ## Q-LoRA config
36
+ - LoRA_r: 16
37
+ - LoRA_alpha: 16
38
+ - LoRA_dropout: 0.05
39
+ - LoRA_target_modules: [gate_proj, up_proj, down_proj]
40
+
41
+ ## Prompt
42
+ - Alpaca template.
43
+
44
+ # **Model Benchmark**
45
+
46
+ ## Open leaderboard
47
+ - Follow up as [link](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
48
+
49
+ | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
50
+ | --- | --- | --- | --- | --- | --- | --- | --- |
51
+ | SOLAR-Platypus-10.7B-v1 | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
52
+ | SOLAR-Platypus-10.7B-v2 | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
53
+ | [upstage/SOLAR-10.7B-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-v1.0) | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
54
+
55
+
56
+ # Implementation Code
57
+ ```python
58
+ ### KO-Platypus
59
+ from transformers import AutoModelForCausalLM, AutoTokenizer
60
+ import torch
61
+
62
+ repo = "kyujinpy/SOLAR-Platypus-10.7B-v1"
63
+ OpenOrca = AutoModelForCausalLM.from_pretrained(
64
+ repo,
65
+ return_dict=True,
66
+ torch_dtype=torch.float16,
67
+ device_map='auto'
68
+ )
69
+ OpenOrca_tokenizer = AutoTokenizer.from_pretrained(repo)
70
+ ```
71
+
72
+ ---