bartowski commited on
Commit
bd7fa7c
1 Parent(s): bc9d6a9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -21
README.md CHANGED
@@ -93,30 +93,36 @@ Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.0.20">turb
93
 
94
  Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions.
95
 
96
- Conversion was done using the default calibration dataset.
97
-
98
- Default arguments used except when the bits per weight is above 6.0, at that point the lm_head layer is quantized at 8 bits per weight instead of the default 6.
99
-
100
  Original model: https://huggingface.co/bigcode/starcoder2-15b-instruct-v0.1
101
 
 
102
 
103
- <a href="https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2/tree/8_0">8.0 bits per weight</a>
104
-
105
- <a href="https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2/tree/6_5">6.5 bits per weight</a>
106
 
107
- <a href="https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2/tree/5_0">5.0 bits per weight</a>
 
108
 
109
- <a href="https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2/tree/4_25">4.25 bits per weight</a>
 
 
110
 
111
- <a href="https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2/tree/3_5">3.5 bits per weight</a>
112
 
 
 
 
 
 
 
 
113
 
114
  ## Download instructions
115
 
116
  With git:
117
 
118
  ```shell
119
- git clone --single-branch --branch 6_5 https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2
120
  ```
121
 
122
  With huggingface hub (credit to TheBloke for instructions):
@@ -125,25 +131,18 @@ With huggingface hub (credit to TheBloke for instructions):
125
  pip3 install huggingface-hub
126
  ```
127
 
128
- To download the `main` (only useful if you only care about measurement.json) branch to a folder called `starcoder2-15b-instruct-v0.1-exl2`:
129
-
130
- ```shell
131
- mkdir starcoder2-15b-instruct-v0.1-exl2
132
- huggingface-cli download bartowski/starcoder2-15b-instruct-v0.1-exl2 --local-dir starcoder2-15b-instruct-v0.1-exl2 --local-dir-use-symlinks False
133
- ```
134
-
135
- To download from a different branch, add the `--revision` parameter:
136
 
137
  Linux:
138
 
139
  ```shell
140
- mkdir starcoder2-15b-instruct-v0.1-exl2-6_5
141
  huggingface-cli download bartowski/starcoder2-15b-instruct-v0.1-exl2 --revision 6_5 --local-dir starcoder2-15b-instruct-v0.1-exl2-6_5 --local-dir-use-symlinks False
142
  ```
143
 
144
  Windows (which apparently doesn't like _ in folders sometimes?):
145
 
146
  ```shell
147
- mkdir starcoder2-15b-instruct-v0.1-exl2-6.5
148
  huggingface-cli download bartowski/starcoder2-15b-instruct-v0.1-exl2 --revision 6_5 --local-dir starcoder2-15b-instruct-v0.1-exl2-6.5 --local-dir-use-symlinks False
149
  ```
 
 
 
93
 
94
  Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions.
95
 
 
 
 
 
96
  Original model: https://huggingface.co/bigcode/starcoder2-15b-instruct-v0.1
97
 
98
+ ## Prompt format
99
 
100
+ ```
101
+ <|endoftext|>You are an exceptionally intelligent coding assistant that consistently delivers accurate and reliable responses to user instructions.
 
102
 
103
+ ### Instruction
104
+ {prompt}
105
 
106
+ ### Response
107
+ <|endoftext|>
108
+ ```
109
 
110
+ ## Available sizes
111
 
112
+ | Branch | Bits | lm_head bits | VRAM (4k) | VRAM (16k) | VRAM (32k) | Description |
113
+ | ----- | ---- | ------- | ------ | ------ | ------ | ------------ |
114
+ | [8_0](https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2/tree/8_0) | 8.0 | 8.0 | 15.8 GB | 16.8 GB | 18.1 GB | Maximum quality that ExLlamaV2 can produce, near unquantized performance. |
115
+ | [6_5](https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2/tree/6_5) | 6.5 | 8.0 | 13.9 GB | 14.9 GB | 16.2 GB | Near unquantized performance at vastly reduced size, **recommended**. |
116
+ | [5_0](https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2/tree/5_0) | 5.0 | 6.0 | 11.0 GB | 12.0 GB | 13.2 GB | Slightly lower quality vs 6.5. |
117
+ | [4_25](https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2/tree/4_25) | 4.25 | 6.0 | 9.5 GB | 10.5 GB | 11.8 GB | GPTQ equivalent bits per weight. |
118
+ | [3_5](https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2/tree/3_5) | 3.5 | 6.0 | 8.1 GB | 9.1 GB | 10.4 GB | Lower quality, not recommended. |
119
 
120
  ## Download instructions
121
 
122
  With git:
123
 
124
  ```shell
125
+ git clone --single-branch --branch 6_5 https://huggingface.co/bartowski/starcoder2-15b-instruct-v0.1-exl2 starcoder2-15b-instruct-v0.1-exl2-6_5
126
  ```
127
 
128
  With huggingface hub (credit to TheBloke for instructions):
 
131
  pip3 install huggingface-hub
132
  ```
133
 
134
+ To download a specific branch, use the `--revision` parameter. For example, to download the 6.5 bpw branch:
 
 
 
 
 
 
 
135
 
136
  Linux:
137
 
138
  ```shell
 
139
  huggingface-cli download bartowski/starcoder2-15b-instruct-v0.1-exl2 --revision 6_5 --local-dir starcoder2-15b-instruct-v0.1-exl2-6_5 --local-dir-use-symlinks False
140
  ```
141
 
142
  Windows (which apparently doesn't like _ in folders sometimes?):
143
 
144
  ```shell
 
145
  huggingface-cli download bartowski/starcoder2-15b-instruct-v0.1-exl2 --revision 6_5 --local-dir starcoder2-15b-instruct-v0.1-exl2-6.5 --local-dir-use-symlinks False
146
  ```
147
+
148
+ Want to support my work? Visit my ko-fi page here: https://ko-fi.com/bartowski