AndreyBest
commited on
Commit
•
4c66790
1
Parent(s):
cf757f8
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,262 @@
|
|
1 |
-
---
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
pipeline_tag: text-generation
|
3 |
+
base_model: ibm-granite/granite-3b-code-base
|
4 |
+
license: apache-2.0
|
5 |
+
datasets:
|
6 |
+
- bigcode/commitpackft
|
7 |
+
- TIGER-Lab/MathInstruct
|
8 |
+
- meta-math/MetaMathQA
|
9 |
+
- glaiveai/glaive-code-assistant-v3
|
10 |
+
- glaive-function-calling-v2
|
11 |
+
- bugdaryan/sql-create-context-instruction
|
12 |
+
- garage-bAInd/Open-Platypus
|
13 |
+
- nvidia/HelpSteer
|
14 |
+
metrics:
|
15 |
+
- code_eval
|
16 |
+
library_name: transformers
|
17 |
+
tags:
|
18 |
+
- code
|
19 |
+
- granite
|
20 |
+
model-index:
|
21 |
+
- name: granite-3b-code-instruct
|
22 |
+
results:
|
23 |
+
- task:
|
24 |
+
type: text-generation
|
25 |
+
dataset:
|
26 |
+
type: bigcode/humanevalpack
|
27 |
+
name: HumanEvalSynthesis(Python)
|
28 |
+
metrics:
|
29 |
+
- name: pass@1
|
30 |
+
type: pass@1
|
31 |
+
value: 51.2
|
32 |
+
veriefied: false
|
33 |
+
- task:
|
34 |
+
type: text-generation
|
35 |
+
dataset:
|
36 |
+
type: bigcode/humanevalpack
|
37 |
+
name: HumanEvalSynthesis(JavaScript)
|
38 |
+
metrics:
|
39 |
+
- name: pass@1
|
40 |
+
type: pass@1
|
41 |
+
value: 43.9
|
42 |
+
veriefied: false
|
43 |
+
- task:
|
44 |
+
type: text-generation
|
45 |
+
dataset:
|
46 |
+
type: bigcode/humanevalpack
|
47 |
+
name: HumanEvalSynthesis(Java)
|
48 |
+
metrics:
|
49 |
+
- name: pass@1
|
50 |
+
type: pass@1
|
51 |
+
value: 41.5
|
52 |
+
veriefied: false
|
53 |
+
- task:
|
54 |
+
type: text-generation
|
55 |
+
dataset:
|
56 |
+
type: bigcode/humanevalpack
|
57 |
+
name: HumanEvalSynthesis(Go)
|
58 |
+
metrics:
|
59 |
+
- name: pass@1
|
60 |
+
type: pass@1
|
61 |
+
value: 31.7
|
62 |
+
veriefied: false
|
63 |
+
- task:
|
64 |
+
type: text-generation
|
65 |
+
dataset:
|
66 |
+
type: bigcode/humanevalpack
|
67 |
+
name: HumanEvalSynthesis(C++)
|
68 |
+
metrics:
|
69 |
+
- name: pass@1
|
70 |
+
type: pass@1
|
71 |
+
value: 40.2
|
72 |
+
veriefied: false
|
73 |
+
- task:
|
74 |
+
type: text-generation
|
75 |
+
dataset:
|
76 |
+
type: bigcode/humanevalpack
|
77 |
+
name: HumanEvalSynthesis(Rust)
|
78 |
+
metrics:
|
79 |
+
- name: pass@1
|
80 |
+
type: pass@1
|
81 |
+
value: 29.3
|
82 |
+
veriefied: false
|
83 |
+
- task:
|
84 |
+
type: text-generation
|
85 |
+
dataset:
|
86 |
+
type: bigcode/humanevalpack
|
87 |
+
name: HumanEvalExplain(Python)
|
88 |
+
metrics:
|
89 |
+
- name: pass@1
|
90 |
+
type: pass@1
|
91 |
+
value: 39.6
|
92 |
+
veriefied: false
|
93 |
+
- task:
|
94 |
+
type: text-generation
|
95 |
+
dataset:
|
96 |
+
type: bigcode/humanevalpack
|
97 |
+
name: HumanEvalExplain(JavaScript)
|
98 |
+
metrics:
|
99 |
+
- name: pass@1
|
100 |
+
type: pass@1
|
101 |
+
value: 26.8
|
102 |
+
veriefied: false
|
103 |
+
- task:
|
104 |
+
type: text-generation
|
105 |
+
dataset:
|
106 |
+
type: bigcode/humanevalpack
|
107 |
+
name: HumanEvalExplain(Java)
|
108 |
+
metrics:
|
109 |
+
- name: pass@1
|
110 |
+
type: pass@1
|
111 |
+
value: 39
|
112 |
+
veriefied: false
|
113 |
+
- task:
|
114 |
+
type: text-generation
|
115 |
+
dataset:
|
116 |
+
type: bigcode/humanevalpack
|
117 |
+
name: HumanEvalExplain(Go)
|
118 |
+
metrics:
|
119 |
+
- name: pass@1
|
120 |
+
type: pass@1
|
121 |
+
value: 14
|
122 |
+
veriefied: false
|
123 |
+
- task:
|
124 |
+
type: text-generation
|
125 |
+
dataset:
|
126 |
+
type: bigcode/humanevalpack
|
127 |
+
name: HumanEvalExplain(C++)
|
128 |
+
metrics:
|
129 |
+
- name: pass@1
|
130 |
+
type: pass@1
|
131 |
+
value: 23.8
|
132 |
+
veriefied: false
|
133 |
+
- task:
|
134 |
+
type: text-generation
|
135 |
+
dataset:
|
136 |
+
type: bigcode/humanevalpack
|
137 |
+
name: HumanEvalExplain(Rust)
|
138 |
+
metrics:
|
139 |
+
- name: pass@1
|
140 |
+
type: pass@1
|
141 |
+
value: 12.8
|
142 |
+
veriefied: false
|
143 |
+
- task:
|
144 |
+
type: text-generation
|
145 |
+
dataset:
|
146 |
+
type: bigcode/humanevalpack
|
147 |
+
name: HumanEvalFix(Python)
|
148 |
+
metrics:
|
149 |
+
- name: pass@1
|
150 |
+
type: pass@1
|
151 |
+
value: 26.8
|
152 |
+
veriefied: false
|
153 |
+
- task:
|
154 |
+
type: text-generation
|
155 |
+
dataset:
|
156 |
+
type: bigcode/humanevalpack
|
157 |
+
name: HumanEvalFix(JavaScript)
|
158 |
+
metrics:
|
159 |
+
- name: pass@1
|
160 |
+
type: pass@1
|
161 |
+
value: 28
|
162 |
+
veriefied: false
|
163 |
+
- task:
|
164 |
+
type: text-generation
|
165 |
+
dataset:
|
166 |
+
type: bigcode/humanevalpack
|
167 |
+
name: HumanEvalFix(Java)
|
168 |
+
metrics:
|
169 |
+
- name: pass@1
|
170 |
+
type: pass@1
|
171 |
+
value: 33.5
|
172 |
+
veriefied: false
|
173 |
+
- task:
|
174 |
+
type: text-generation
|
175 |
+
dataset:
|
176 |
+
type: bigcode/humanevalpack
|
177 |
+
name: HumanEvalFix(Go)
|
178 |
+
metrics:
|
179 |
+
- name: pass@1
|
180 |
+
type: pass@1
|
181 |
+
value: 27.4
|
182 |
+
veriefied: false
|
183 |
+
- task:
|
184 |
+
type: text-generation
|
185 |
+
dataset:
|
186 |
+
type: bigcode/humanevalpack
|
187 |
+
name: HumanEvalFix(C++)
|
188 |
+
metrics:
|
189 |
+
- name: pass@1
|
190 |
+
type: pass@1
|
191 |
+
value: 31.7
|
192 |
+
veriefied: false
|
193 |
+
- task:
|
194 |
+
type: text-generation
|
195 |
+
dataset:
|
196 |
+
type: bigcode/humanevalpack
|
197 |
+
name: HumanEvalFix(Rust)
|
198 |
+
metrics:
|
199 |
+
- name: pass@1
|
200 |
+
type: pass@1
|
201 |
+
value: 16.5
|
202 |
+
veriefied: false
|
203 |
+
---
|
204 |
+
|
205 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64a28db2f1968b7d7f357182/kaudiTlvRQBA5NSeq4BbM.png)
|
206 |
+
*This model was quantized by [SanctumAI](https://sanctum.ai). To leave feedback, join our community in [Discord](https://discord.gg/7ZNE78HJKh).*
|
207 |
+
|
208 |
+
# Granite 3B Code Instruct GGUF
|
209 |
+
|
210 |
+
**Model creator:** [ibm-granite](https://huggingface.co/ibm-granite)<br>
|
211 |
+
**Original model**: [granite-3b-code-instruct](https://huggingface.co/ibm-granite/granite-3b-code-instruct)<br>
|
212 |
+
|
213 |
+
## Model Summary:
|
214 |
+
|
215 |
+
**Granite-3B-Code-Instruct** is a 3B parameter model fine tuned from *Granite-3B-Code-Base* on a combination of **permissively licensed** instruction data to enhance instruction following capabilities including logical reasoning and problem-solving skills.
|
216 |
+
|
217 |
+
- **Developers:** IBM Research
|
218 |
+
- **GitHub Repository:** [ibm-granite/granite-code-models](https://github.com/ibm-granite/granite-code-models)
|
219 |
+
- **Paper:** [Granite Code Models: A Family of Open Foundation Models for Code Intelligence](https://arxiv.org/abs/2405.04324)
|
220 |
+
- **Release Date**: May 6th, 2024
|
221 |
+
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0).
|
222 |
+
|
223 |
+
## Prompt Template:
|
224 |
+
|
225 |
+
If you're using Sanctum app, simply use `IBM Granite Code` model preset.
|
226 |
+
|
227 |
+
Prompt template:
|
228 |
+
|
229 |
+
```
|
230 |
+
System:
|
231 |
+
{system_prompt}
|
232 |
+
Question:
|
233 |
+
{prompt}
|
234 |
+
Answer:
|
235 |
+
|
236 |
+
```
|
237 |
+
|
238 |
+
## Hardware Requirements Estimate
|
239 |
+
|
240 |
+
| Name | Quant method | Size | Memory (RAM, vRAM) required (for full context of 32k tokens) |
|
241 |
+
| ---- | ---- | ---- | ---- |
|
242 |
+
| [granite-3b-code-instruct.Q2_K.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q2_K.gguf) | Q2_K | 1.34 GB | 4.68 GB |
|
243 |
+
| [granite-3b-code-instruct.Q3_K_S.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q3_K_S.gguf) | Q3_K_S | 1.55 GB | 7.19 GB |
|
244 |
+
| [granite-3b-code-instruct.Q3_K_M.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q3_K_M.gguf) | Q3_K_M | 1.73 GB | 7.52 GB |
|
245 |
+
| [granite-3b-code-instruct.Q3_K_L.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q3_K_L.gguf) | Q3_K_L | 1.88 GB | 7.80 GB |
|
246 |
+
| [granite-3b-code-instruct.Q4_0.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q4_0.gguf) | Q4_0 | 2.00 GB | 8.07 GB |
|
247 |
+
| [granite-3b-code-instruct.Q4_K_S.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q4_K_S.gguf) | Q4_K_S | 2.01 GB | 8.10 GB |
|
248 |
+
| [granite-3b-code-instruct.Q4_K_M.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q4_K_M.gguf) | Q4_K_M | 2.13 GB | 8.31 GB |
|
249 |
+
| [granite-3b-code-instruct.Q4_K.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q4_K.gguf) | Q4_K | 2.13 GB | 8.31 GB |
|
250 |
+
| [granite-3b-code-instruct.Q4_1.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q4_1.gguf) | Q4_1 | 2.21 GB | 8.48 GB |
|
251 |
+
| [granite-3b-code-instruct.Q5_0.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q5_0.gguf) | Q5_0 | 2.42 GB | 8.90 GB |
|
252 |
+
| [granite-3b-code-instruct.Q5_K_S.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q5_K_S.gguf) | Q5_K_S | 2.42 GB | 8.90 GB |
|
253 |
+
| [granite-3b-code-instruct.Q5_K_M.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q5_K_M.gguf) | Q5_K_M | 2.49 GB | 9.02 GB |
|
254 |
+
| [granite-3b-code-instruct.Q5_K.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q5_K.gguf) | Q5_K | 2.49 GB | 9.02 GB |
|
255 |
+
| [granite-3b-code-instruct.Q5_1.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q5_1.gguf) | Q5_1 | 2.63 GB | 9.31 GB |
|
256 |
+
| [granite-3b-code-instruct.Q6_K.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q6_K.gguf) | Q6_K | 2.86 GB | 9.78 GB |
|
257 |
+
| [granite-3b-code-instruct.Q8_0.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.Q8_0.gguf) | Q8_0 | 3.71 GB | 11.41 GB |
|
258 |
+
| [granite-3b-code-instruct.f16.gguf](https://huggingface.co/SanctumAI/granite-3b-code-instruct-GGUF/blob/main/granite-3b-code-instruct.f16.gguf) | f16 | 6.97 GB | 4.68 GB |
|
259 |
+
|
260 |
+
## Disclaimer
|
261 |
+
|
262 |
+
Sanctum is not the creator, originator, or owner of any Model featured in the Models section of the Sanctum application. Each Model is created and provided by third parties. Sanctum does not endorse, support, represent or guarantee the completeness, truthfulness, accuracy, or reliability of any Model listed there. You understand that supported Models can produce content that might be offensive, harmful, inaccurate or otherwise inappropriate, or deceptive. Each Model is the sole responsibility of the person or entity who originated such Model. Sanctum may not monitor or control the Models supported and cannot, and does not, take responsibility for any such Model. Sanctum disclaims all warranties or guarantees about the accuracy, reliability or benefits of the Models. Sanctum further disclaims any warranty that the Model will meet your requirements, be secure, uninterrupted or available at any time or location, or error-free, viruses-free, or that any errors will be corrected, or otherwise. You will be solely responsible for any damage resulting from your use of or access to the Models, your downloading of any Model, or use of any other Model provided by or through Sanctum.
|