cgus commited on
Commit
00227a4
1 Parent(s): ed267f4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -3
README.md CHANGED
@@ -1,13 +1,36 @@
1
  ---
2
  base_model:
3
- - NousResearch/Hermes-3-Llama-3.1-8B
4
- library_name: transformers
5
  tags:
6
  - mergekit
7
  - merge
8
 
9
  ---
10
- # 🪽 Hermes-3-Llama-3.1-8B-lorablated
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
 
12
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/4Hbw5n68jKUSBQeTqQIeT.png)
13
  <center>70B version: <a href="https://huggingface.co/mlabonne/Hermes-3-Llama-3.1-70B-lorablated/"><i>mlabonne/Hermes-3-Llama-3.1-70B-lorablated</i></a></center>
 
1
  ---
2
  base_model:
3
+ - mlabonne/Hermes-3-Llama-3.1-8B-lorablated
4
+ license: llama3
5
  tags:
6
  - mergekit
7
  - merge
8
 
9
  ---
10
+ # Hermes-3-Llama-3.1-8B-lorablated-exl2
11
+ Model: [Hermes-3-Llama-3.1-8B-lorablated](https://huggingface.co/mlabonne/Hermes-3-Llama-3.1-8B-lorablated)
12
+ Created by: [mlabonne](https://huggingface.co/mlabonne)
13
+ Based on: [Hermes-3-Llama-3.1-8B](https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B)
14
+
15
+ ## Quants
16
+ [4bpw h6](https://huggingface.co/cgus/Hermes-3-Llama-3.1-8B-lorablated-exl2/tree/main)
17
+ [4.5bpw h6](https://huggingface.co/cgus/Hermes-3-Llama-3.1-8B-lorablated-exl2/tree/4.5bpw-h6)
18
+ [5bpw h6](https://huggingface.co/cgus/Hermes-3-Llama-3.1-8B-lorablated-exl2/tree/5bpw-h6)
19
+ [6bpw h6](https://huggingface.co/cgus/Hermes-3-Llama-3.1-8B-lorablated-exl2/tree/6bpw-h6)
20
+ [8bpw h8](https://huggingface.co/cgus/Hermes-3-Llama-3.1-8B-lorablated-exl2/tree/8bpw-h8)
21
+
22
+ ## Quantization notes
23
+ Made with Exllamav2 0.1.8 with the default dataset.
24
+ I'm not sure how well it works with Text-Generation-WebUI considering that this model uses some unusual RoPE mechanics and I have no idea how TGW handles it.
25
+ For some reason this model worked extremely slow with my TGW install but was perfectly fine with TabbyAPI.
26
+
27
+ ## How to run
28
+ I recommend using TabbyAPI for this model. The model requires a decent Nvidia RTX card on Windows/Linux or a decent AMD GPU on Linux.
29
+ It requires to be fully loaded in GPU to work, so if your GPU has too small VRAM you should use [GGUF version](https://huggingface.co/mlabonne/Hermes-3-Llama-3.1-8B-lorablated-GGUF) instead.
30
+ If you have Nvidia GTX card you should also use GGUF instead.
31
+
32
+ # Orignal model card
33
+ # Hermes-3-Llama-3.1-8B-lorablated
34
 
35
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/4Hbw5n68jKUSBQeTqQIeT.png)
36
  <center>70B version: <a href="https://huggingface.co/mlabonne/Hermes-3-Llama-3.1-70B-lorablated/"><i>mlabonne/Hermes-3-Llama-3.1-70B-lorablated</i></a></center>