morriszms commited on
Commit
b827da9
1 Parent(s): f310023

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,16 +1,16 @@
1
  ---
2
- language:
3
- - en
4
- library_name: transformers
5
  license: gemma
 
 
 
 
 
 
 
 
6
  tags:
7
- - unsloth
8
- - transformers
9
- - gemma2
10
- - gemma
11
  - TensorBlock
12
  - GGUF
13
- base_model: unsloth/gemma-2-9b
14
  ---
15
 
16
  <div style="width: auto; margin-left: auto; margin-right: auto">
@@ -24,13 +24,12 @@ base_model: unsloth/gemma-2-9b
24
  </div>
25
  </div>
26
 
27
- ## unsloth/gemma-2-9b - GGUF
28
 
29
- This repo contains GGUF format model files for [unsloth/gemma-2-9b](https://huggingface.co/unsloth/gemma-2-9b).
30
 
31
  The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
32
 
33
-
34
  <div style="text-align: left; margin: 20px 0;">
35
  <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
36
  Run them on the TensorBlock client using your local machine ↗
@@ -39,7 +38,6 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
39
 
40
  ## Prompt template
41
 
42
-
43
  ```
44
 
45
  ```
@@ -48,18 +46,18 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
48
 
49
  | Filename | Quant type | File Size | Description |
50
  | -------- | ---------- | --------- | ----------- |
51
- | [gemma-2-9b-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q2_K.gguf) | Q2_K | 3.544 GB | smallest, significant quality loss - not recommended for most purposes |
52
- | [gemma-2-9b-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q3_K_S.gguf) | Q3_K_S | 4.040 GB | very small, high quality loss |
53
- | [gemma-2-9b-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q3_K_M.gguf) | Q3_K_M | 4.435 GB | very small, high quality loss |
54
- | [gemma-2-9b-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q3_K_L.gguf) | Q3_K_L | 4.780 GB | small, substantial quality loss |
55
- | [gemma-2-9b-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q4_0.gguf) | Q4_0 | 5.069 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
56
- | [gemma-2-9b-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q4_K_S.gguf) | Q4_K_S | 5.103 GB | small, greater quality loss |
57
- | [gemma-2-9b-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q4_K_M.gguf) | Q4_K_M | 5.365 GB | medium, balanced quality - recommended |
58
- | [gemma-2-9b-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q5_0.gguf) | Q5_0 | 6.038 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
59
- | [gemma-2-9b-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q5_K_S.gguf) | Q5_K_S | 6.038 GB | large, low quality loss - recommended |
60
- | [gemma-2-9b-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q5_K_M.gguf) | Q5_K_M | 6.191 GB | large, very low quality loss - recommended |
61
- | [gemma-2-9b-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q6_K.gguf) | Q6_K | 7.068 GB | very large, extremely low quality loss |
62
- | [gemma-2-9b-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q8_0.gguf) | Q8_0 | 9.152 GB | very large, extremely low quality loss - not recommended |
63
 
64
 
65
  ## Downloading instruction
 
1
  ---
 
 
 
2
  license: gemma
3
+ library_name: transformers
4
+ pipeline_tag: text-generation
5
+ extra_gated_heading: Access Gemma on Hugging Face
6
+ extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and
7
+ agree to Google’s usage license. To do this, please ensure you’re logged in to Hugging
8
+ Face and click below. Requests are processed immediately.
9
+ extra_gated_button_content: Acknowledge license
10
+ base_model: google/gemma-2-9b
11
  tags:
 
 
 
 
12
  - TensorBlock
13
  - GGUF
 
14
  ---
15
 
16
  <div style="width: auto; margin-left: auto; margin-right: auto">
 
24
  </div>
25
  </div>
26
 
27
+ ## google/gemma-2-9b - GGUF
28
 
29
+ This repo contains GGUF format model files for [google/gemma-2-9b](https://huggingface.co/google/gemma-2-9b).
30
 
31
  The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
32
 
 
33
  <div style="text-align: left; margin: 20px 0;">
34
  <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
35
  Run them on the TensorBlock client using your local machine ↗
 
38
 
39
  ## Prompt template
40
 
 
41
  ```
42
 
43
  ```
 
46
 
47
  | Filename | Quant type | File Size | Description |
48
  | -------- | ---------- | --------- | ----------- |
49
+ | [gemma-2-9b-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q2_K.gguf) | Q2_K | 3.805 GB | smallest, significant quality loss - not recommended for most purposes |
50
+ | [gemma-2-9b-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q3_K_S.gguf) | Q3_K_S | 4.338 GB | very small, high quality loss |
51
+ | [gemma-2-9b-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q3_K_M.gguf) | Q3_K_M | 4.762 GB | very small, high quality loss |
52
+ | [gemma-2-9b-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q3_K_L.gguf) | Q3_K_L | 5.132 GB | small, substantial quality loss |
53
+ | [gemma-2-9b-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q4_0.gguf) | Q4_0 | 5.443 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
54
+ | [gemma-2-9b-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q4_K_S.gguf) | Q4_K_S | 5.479 GB | small, greater quality loss |
55
+ | [gemma-2-9b-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q4_K_M.gguf) | Q4_K_M | 5.761 GB | medium, balanced quality - recommended |
56
+ | [gemma-2-9b-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q5_0.gguf) | Q5_0 | 6.484 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
57
+ | [gemma-2-9b-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q5_K_S.gguf) | Q5_K_S | 6.484 GB | large, low quality loss - recommended |
58
+ | [gemma-2-9b-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q5_K_M.gguf) | Q5_K_M | 6.647 GB | large, very low quality loss - recommended |
59
+ | [gemma-2-9b-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q6_K.gguf) | Q6_K | 7.589 GB | very large, extremely low quality loss |
60
+ | [gemma-2-9b-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-GGUF/blob/main/gemma-2-9b-Q8_0.gguf) | Q8_0 | 9.827 GB | very large, extremely low quality loss - not recommended |
61
 
62
 
63
  ## Downloading instruction
gemma-2-9b-Q2_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6af6d1a8a93de3f6361af9ff8bfd8eb063d85539b9c6762ff23ffeec7f83d5af
3
- size 3805397376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0f0f62ff0b6cd8f2af7763f6035d3af9b7d5999f13a29bd49cbcb6e264669b1
3
+ size 3805397280
gemma-2-9b-Q3_K_L.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5620d6a07c5797eca9b82ff231a8136be2f8b37bef4973e972053517988bcbed
3
- size 5132452224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cb628776cb02bb5cfa80d60fe8b6543d3656e38d80edc6f670cbcca007b3334
3
+ size 5132452128
gemma-2-9b-Q3_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c07669932686f109abccbe89c740a52b4eaee1e7c59e44ca9c87069aac05bf5
3
- size 4761780608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d8639bdf1db18d7ccf9b6e6a78806dd839bbed5b84bdb6b6d8906b8ca1f29fa
3
+ size 4761780512
gemma-2-9b-Q3_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b50703e49fce96d8dd39c735e51acd36946a0be8161ab4c5b89c6865fda9e0b2
3
- size 4337664384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5484c2b6db099e00fa5d682841080772e7c1716de6b0a5b442e1f252c979a6d5
3
+ size 4337664288
gemma-2-9b-Q4_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2d98335b9b0003db0ceac3eb94ff69b8787abda1aea5367b60430574bf5ff32
3
- size 5443142016
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7727d251eb83f63988e5dc7c5d61494b7f325cae37eb7fa6e5017b2ae7febf7
3
+ size 5443141920
gemma-2-9b-Q4_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b609d854cba451805e8e74e6c8e6a72f26dcc1e1c6deda72d4a1207a83dfcc8
3
- size 5761057152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2ce0f5b179ad209d0f7461d15a81bd1db975b452247290da5e72508128f8139
3
+ size 5761057056
gemma-2-9b-Q4_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e587eba863465183e14b5f05f755b7457973b722b1d5fbf51071e3117c4f7b6
3
- size 5478924672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4440de062603ee96e5bec7ae46b68edacbb08b62f897bf7ef9c3168e544c5a82
3
+ size 5478924576
gemma-2-9b-Q5_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:683109a1123a175da4500ebcdc4ff29ed68b1ab60ad25c2cbd5077aaadf199b8
3
- size 6483591552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad2438dd6236528aa9e6da099e380dab20b806432625c53a5728e80725dac2f8
3
+ size 6483591456
gemma-2-9b-Q5_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c79ee86231c3c6997692b789fdabf4a77f813c87c30d5c46766a6bdc05c85940
3
- size 6647366016
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d43c0777c38d6eb835c26c14aeda656df2c9ad31d6a6255f3f479520b0d4535
3
+ size 6647365920
gemma-2-9b-Q5_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:decfe2431ad3996aa7cb574cf83337bfef7c1f733ad8a578affae664b3fa1051
3
- size 6483591552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c166542809a3d054b9632b8d2f770006da574c699c03f0ba64519c42cf37a86
3
+ size 6483591456
gemma-2-9b-Q6_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:122ec01c79d2288a849b9b0e6f10019ff75c070d97067a41af3475839b841379
3
- size 7589069184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67a8fd5eb1507eb966b33905c145e8f3b9bfea9dad0cdeedad3e07d37892070a
3
+ size 7589069088
gemma-2-9b-Q8_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:417157e97f370b176f6c7c092ef7e664acdbc574debad851c2c02e008d4b6afd
3
- size 9827148160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93f7d9c9143fabe4ba1f9811cfa39ace3c889956bda6e4ace2115fe4646fee32
3
+ size 9827148064