AIronMind commited on
Commit
71ae2bf
1 Parent(s): 6101faf

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +187 -0
README.md ADDED
@@ -0,0 +1,187 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - zh
5
+ - id
6
+ - th
7
+ - vi
8
+ - ms
9
+ - lo
10
+ datasets:
11
+ - cerebras/SlimPajama-627B
12
+ - Skywork/SkyPile-150B
13
+ - allenai/MADLAD-400
14
+ - cc100
15
+ tags:
16
+ - multilingual
17
+ - sea
18
+ - sailor
19
+ - llama-cpp
20
+ - gguf-my-repo
21
+ license: apache-2.0
22
+ base_model: sail/Sailor-7B
23
+ model-index:
24
+ - name: Sailor-7B
25
+ results:
26
+ - task:
27
+ type: text-generation
28
+ dataset:
29
+ name: XQuAD-Thai
30
+ type: XQuAD-Thai
31
+ metrics:
32
+ - type: EM (3-Shot)
33
+ value: 57.88
34
+ name: EM (3-Shot)
35
+ - type: F1 (3-Shot)
36
+ value: 71.06
37
+ name: F1 (3-Shot)
38
+ - task:
39
+ type: text-generation
40
+ dataset:
41
+ name: TyDiQA-Indonesian
42
+ type: TyDiQA-Indonesian
43
+ metrics:
44
+ - type: EM (3-Shot)
45
+ value: 60.53
46
+ name: EM (3-Shot)
47
+ - type: F1 (3-Shot)
48
+ value: 75.42
49
+ name: F1 (3-Shot)
50
+ - task:
51
+ type: text-generation
52
+ dataset:
53
+ name: XQuAD-Vietnamese
54
+ type: XQuAD-Vietnamese
55
+ metrics:
56
+ - type: EM (3-Shot)
57
+ value: 53.81
58
+ name: EM (3-Shot)
59
+ - type: F1 (3-Shot)
60
+ value: 74.62
61
+ name: F1 (3-Shot)
62
+ - task:
63
+ type: text-generation
64
+ dataset:
65
+ name: XCOPA-Thai
66
+ type: XCOPA-Thai
67
+ metrics:
68
+ - type: EM (3-Shot)
69
+ value: 59.0
70
+ name: EM (3-Shot)
71
+ - task:
72
+ type: text-generation
73
+ dataset:
74
+ name: XCOPA-Indonesian
75
+ type: XCOPA-Indonesian
76
+ metrics:
77
+ - type: EM (3-Shot)
78
+ value: 72.2
79
+ name: EM (3-Shot)
80
+ - task:
81
+ type: text-generation
82
+ dataset:
83
+ name: XCOPA-Vietnamese
84
+ type: XCOPA-Vietnamese
85
+ metrics:
86
+ - type: EM (3-Shot)
87
+ value: 72.2
88
+ name: EM (3-Shot)
89
+ - task:
90
+ type: text-generation
91
+ dataset:
92
+ name: M3Exam-Thai
93
+ type: M3Exam-Thai
94
+ metrics:
95
+ - type: EM (3-Shot)
96
+ value: 30.0
97
+ name: EM (3-Shot)
98
+ - task:
99
+ type: text-generation
100
+ dataset:
101
+ name: M3Exam-Indonesian
102
+ type: M3Exam-Indonesian
103
+ metrics:
104
+ - type: EM (3-Shot)
105
+ value: 32.88
106
+ name: EM (3-Shot)
107
+ - task:
108
+ type: text-generation
109
+ dataset:
110
+ name: M3Exam-Vietnamese
111
+ type: M3Exam-Vietnamese
112
+ metrics:
113
+ - type: EM (3-Shot)
114
+ value: 44.1
115
+ name: EM (3-Shot)
116
+ - task:
117
+ type: text-generation
118
+ dataset:
119
+ name: BELEBELE-Thai
120
+ type: BELEBELE-Thai
121
+ metrics:
122
+ - type: EM (3-Shot)
123
+ value: 41.56
124
+ name: EM (3-Shot)
125
+ - task:
126
+ type: text-generation
127
+ dataset:
128
+ name: BELEBELE-Indonesian
129
+ type: BELEBELE-Indonesian
130
+ metrics:
131
+ - type: EM (3-Shot)
132
+ value: 44.33
133
+ name: EM (3-Shot)
134
+ - task:
135
+ type: text-generation
136
+ dataset:
137
+ name: BELEBELE-Vietnamese
138
+ type: BELEBELE-Vietnamese
139
+ metrics:
140
+ - type: EM (3-Shot)
141
+ value: 45.33
142
+ name: EM (3-Shot)
143
+ ---
144
+
145
+ # AIronMind/Sailor-7B-Q4_K_M-GGUF
146
+ This model was converted to GGUF format from [`sail/Sailor-7B`](https://huggingface.co/sail/Sailor-7B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
147
+ Refer to the [original model card](https://huggingface.co/sail/Sailor-7B) for more details on the model.
148
+
149
+ ## Use with llama.cpp
150
+ Install llama.cpp through brew (works on Mac and Linux)
151
+
152
+ ```bash
153
+ brew install llama.cpp
154
+
155
+ ```
156
+ Invoke the llama.cpp server or the CLI.
157
+
158
+ ### CLI:
159
+ ```bash
160
+ llama-cli --hf-repo AIronMind/Sailor-7B-Q4_K_M-GGUF --hf-file sailor-7b-q4_k_m.gguf -p "The meaning to life and the universe is"
161
+ ```
162
+
163
+ ### Server:
164
+ ```bash
165
+ llama-server --hf-repo AIronMind/Sailor-7B-Q4_K_M-GGUF --hf-file sailor-7b-q4_k_m.gguf -c 2048
166
+ ```
167
+
168
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
169
+
170
+ Step 1: Clone llama.cpp from GitHub.
171
+ ```
172
+ git clone https://github.com/ggerganov/llama.cpp
173
+ ```
174
+
175
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
176
+ ```
177
+ cd llama.cpp && LLAMA_CURL=1 make
178
+ ```
179
+
180
+ Step 3: Run inference through the main binary.
181
+ ```
182
+ ./llama-cli --hf-repo AIronMind/Sailor-7B-Q4_K_M-GGUF --hf-file sailor-7b-q4_k_m.gguf -p "The meaning to life and the universe is"
183
+ ```
184
+ or
185
+ ```
186
+ ./llama-server --hf-repo AIronMind/Sailor-7B-Q4_K_M-GGUF --hf-file sailor-7b-q4_k_m.gguf -c 2048
187
+ ```