teemperor commited on
Commit
e3ed355
1 Parent(s): e0a6152

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +115 -0
README.md ADDED
@@ -0,0 +1,115 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: bigcode/starcoder2-15b
3
+ datasets:
4
+ - bigcode/the-stack-v2-train
5
+ library_name: transformers
6
+ license: bigcode-openrail-m
7
+ pipeline_tag: text-generation
8
+ tags:
9
+ - code
10
+ - llama-cpp
11
+ - gguf-my-repo
12
+ inference:
13
+ parameters:
14
+ temperature: 0.2
15
+ top_p: 0.95
16
+ widget:
17
+ - text: 'def print_hello_world():'
18
+ example_title: Hello world
19
+ group: Python
20
+ model-index:
21
+ - name: starcoder2-15b
22
+ results:
23
+ - task:
24
+ type: text-generation
25
+ dataset:
26
+ name: CruxEval-I
27
+ type: cruxeval-i
28
+ metrics:
29
+ - type: pass@1
30
+ value: 48.1
31
+ - task:
32
+ type: text-generation
33
+ dataset:
34
+ name: DS-1000
35
+ type: ds-1000
36
+ metrics:
37
+ - type: pass@1
38
+ value: 33.8
39
+ - task:
40
+ type: text-generation
41
+ dataset:
42
+ name: GSM8K (PAL)
43
+ type: gsm8k-pal
44
+ metrics:
45
+ - type: accuracy
46
+ value: 65.1
47
+ - task:
48
+ type: text-generation
49
+ dataset:
50
+ name: HumanEval+
51
+ type: humanevalplus
52
+ metrics:
53
+ - type: pass@1
54
+ value: 37.8
55
+ - task:
56
+ type: text-generation
57
+ dataset:
58
+ name: HumanEval
59
+ type: humaneval
60
+ metrics:
61
+ - type: pass@1
62
+ value: 46.3
63
+ - task:
64
+ type: text-generation
65
+ dataset:
66
+ name: RepoBench-v1.1
67
+ type: repobench-v1.1
68
+ metrics:
69
+ - type: edit-smiliarity
70
+ value: 74.08
71
+ ---
72
+
73
+ # teemperor/starcoder2-15b-Q6_K-GGUF
74
+ This model was converted to GGUF format from [`bigcode/starcoder2-15b`](https://huggingface.co/bigcode/starcoder2-15b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
75
+ Refer to the [original model card](https://huggingface.co/bigcode/starcoder2-15b) for more details on the model.
76
+
77
+ ## Use with llama.cpp
78
+ Install llama.cpp through brew (works on Mac and Linux)
79
+
80
+ ```bash
81
+ brew install llama.cpp
82
+
83
+ ```
84
+ Invoke the llama.cpp server or the CLI.
85
+
86
+ ### CLI:
87
+ ```bash
88
+ llama-cli --hf-repo teemperor/starcoder2-15b-Q6_K-GGUF --hf-file starcoder2-15b-q6_k.gguf -p "The meaning to life and the universe is"
89
+ ```
90
+
91
+ ### Server:
92
+ ```bash
93
+ llama-server --hf-repo teemperor/starcoder2-15b-Q6_K-GGUF --hf-file starcoder2-15b-q6_k.gguf -c 2048
94
+ ```
95
+
96
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
97
+
98
+ Step 1: Clone llama.cpp from GitHub.
99
+ ```
100
+ git clone https://github.com/ggerganov/llama.cpp
101
+ ```
102
+
103
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
104
+ ```
105
+ cd llama.cpp && LLAMA_CURL=1 make
106
+ ```
107
+
108
+ Step 3: Run inference through the main binary.
109
+ ```
110
+ ./llama-cli --hf-repo teemperor/starcoder2-15b-Q6_K-GGUF --hf-file starcoder2-15b-q6_k.gguf -p "The meaning to life and the universe is"
111
+ ```
112
+ or
113
+ ```
114
+ ./llama-server --hf-repo teemperor/starcoder2-15b-Q6_K-GGUF --hf-file starcoder2-15b-q6_k.gguf -c 2048
115
+ ```