bartowski commited on
Commit
4e4f634
1 Parent(s): e083048

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +8 -7
README.md CHANGED
@@ -1,14 +1,11 @@
1
  ---
2
- base_model: aws-prototyping/MegaBeam-Mistral-7B-512k
3
- license: apache-2.0
4
- pipeline_tag: text-generation
5
  quantized_by: bartowski
6
- inference: false
7
  ---
8
 
9
  ## Llamacpp imatrix Quantizations of MegaBeam-Mistral-7B-512k
10
 
11
- Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b3496">b3496</a> for quantization.
12
 
13
  Original model: https://huggingface.co/aws-prototyping/MegaBeam-Mistral-7B-512k
14
 
@@ -19,16 +16,20 @@ Run them in [LM Studio](https://lmstudio.ai/)
19
  ## Prompt format
20
 
21
  ```
22
- <s> [INST] {prompt} [/INST]</s>
23
  ```
24
 
25
  Note that this model does not support a System prompt.
26
 
 
 
 
 
27
  ## Download a file (not the whole branch) from below:
28
 
29
  | Filename | Quant type | File Size | Split | Description |
30
  | -------- | ---------- | --------- | ----- | ----------- |
31
- | [MegaBeam-Mistral-7B-512k-f32.gguf](https://huggingface.co/bartowski/MegaBeam-Mistral-7B-512k-GGUF/blob/main/MegaBeam-Mistral-7B-512k-f32.gguf) | f32 | 28.97GB | false | Full F32 weights. |
32
  | [MegaBeam-Mistral-7B-512k-Q8_0.gguf](https://huggingface.co/bartowski/MegaBeam-Mistral-7B-512k-GGUF/blob/main/MegaBeam-Mistral-7B-512k-Q8_0.gguf) | Q8_0 | 7.70GB | false | Extremely high quality, generally unneeded but max available quant. |
33
  | [MegaBeam-Mistral-7B-512k-Q6_K_L.gguf](https://huggingface.co/bartowski/MegaBeam-Mistral-7B-512k-GGUF/blob/main/MegaBeam-Mistral-7B-512k-Q6_K_L.gguf) | Q6_K_L | 6.01GB | false | Uses Q8_0 for embed and output weights. Very high quality, near perfect, *recommended*. |
34
  | [MegaBeam-Mistral-7B-512k-Q6_K.gguf](https://huggingface.co/bartowski/MegaBeam-Mistral-7B-512k-GGUF/blob/main/MegaBeam-Mistral-7B-512k-Q6_K.gguf) | Q6_K | 5.94GB | false | Very high quality, near perfect, *recommended*. |
 
1
  ---
 
 
 
2
  quantized_by: bartowski
3
+ pipeline_tag: text-generation
4
  ---
5
 
6
  ## Llamacpp imatrix Quantizations of MegaBeam-Mistral-7B-512k
7
 
8
+ Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b3600">b3600</a> for quantization.
9
 
10
  Original model: https://huggingface.co/aws-prototyping/MegaBeam-Mistral-7B-512k
11
 
 
16
  ## Prompt format
17
 
18
  ```
19
+ <s> [INST] {prompt} [/INST]
20
  ```
21
 
22
  Note that this model does not support a System prompt.
23
 
24
+ ## What's new:
25
+
26
+ Model updated for "improved user experience" and fixing repetition issues
27
+
28
  ## Download a file (not the whole branch) from below:
29
 
30
  | Filename | Quant type | File Size | Split | Description |
31
  | -------- | ---------- | --------- | ----- | ----------- |
32
+ | [MegaBeam-Mistral-7B-512k-f16.gguf](https://huggingface.co/bartowski/MegaBeam-Mistral-7B-512k-GGUF/blob/main/MegaBeam-Mistral-7B-512k-f16.gguf) | f16 | 14.48GB | false | Full F16 weights. |
33
  | [MegaBeam-Mistral-7B-512k-Q8_0.gguf](https://huggingface.co/bartowski/MegaBeam-Mistral-7B-512k-GGUF/blob/main/MegaBeam-Mistral-7B-512k-Q8_0.gguf) | Q8_0 | 7.70GB | false | Extremely high quality, generally unneeded but max available quant. |
34
  | [MegaBeam-Mistral-7B-512k-Q6_K_L.gguf](https://huggingface.co/bartowski/MegaBeam-Mistral-7B-512k-GGUF/blob/main/MegaBeam-Mistral-7B-512k-Q6_K_L.gguf) | Q6_K_L | 6.01GB | false | Uses Q8_0 for embed and output weights. Very high quality, near perfect, *recommended*. |
35
  | [MegaBeam-Mistral-7B-512k-Q6_K.gguf](https://huggingface.co/bartowski/MegaBeam-Mistral-7B-512k-GGUF/blob/main/MegaBeam-Mistral-7B-512k-Q6_K.gguf) | Q6_K | 5.94GB | false | Very high quality, near perfect, *recommended*. |