hfl
/

hfl-rc commited on
Commit
748321d
1 Parent(s): ce22a4e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -2
README.md CHANGED
@@ -1,6 +1,41 @@
1
  ---
2
  license: apache-2.0
 
 
 
3
  ---
4
- Work-in-Progress. No content right now.
5
 
6
- Try v2: https://huggingface.co/hfl/llama-3-chinese-8b-instruct-v2-gguf
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - zh
5
+ - en
6
  ---
 
7
 
8
+ # Llama-3-Chinese-8B-Instruct-v3-GGUF
9
+
10
+ <p align="center">
11
+ <a href="https://github.com/ymcui/Chinese-LLaMA-Alpaca-3"><img src="https://ymcui.com/images/chinese-llama-alpaca-3-banner.png" width="600"/></a>
12
+ </p>
13
+
14
+ This repository contains **Llama-3-Chinese-8B-Instruct-v3-GGUF** (llama.cpp/ollama/tgw, etc. compatible), which is the quantized version of [Llama-3-Chinese-8B-Instruct-v3](https://huggingface.co/hfl/llama-3-chinese-8b-instruct-v3).
15
+
16
+ **Note: this is an instruction (chat) model, which can be used for conversation, QA, etc.**
17
+
18
+ Further details (performance, usage, etc.) should refer to GitHub project page: https://github.com/ymcui/Chinese-LLaMA-Alpaca-3
19
+
20
+ ## Performance
21
+
22
+ Metric: PPL, lower is better
23
+
24
+ *Note: Unless constrained by memory, we suggest using Q8_0 or Q6_K for better performance.*
25
+
26
+ | Quant | Size | PPL |
27
+ | :---: | -------: | ------------------: |
28
+ | Q2_K | 2.96 GB | 10.0534 +/- 0.13135 |
29
+ | Q3_K | 3.74 GB | 6.3295 +/- 0.07816 |
30
+ | Q4_0 | 4.34 GB | 6.3200 +/- 0.07893 |
31
+ | Q4_K | 4.58 GB | 6.0042 +/- 0.07431 |
32
+ | Q5_0 | 5.21 GB | 6.0437 +/- 0.07526 |
33
+ | Q5_K | 5.34 GB | 5.9484 +/- 0.07399 |
34
+ | Q6_K | 6.14 GB | 5.9469 +/- 0.07404 |
35
+ | Q8_0 | 7.95 GB | 5.8933 +/- 0.07305 |
36
+ | F16 | 14.97 GB | 5.8902 +/- 0.07303 |
37
+
38
+ ## Others
39
+
40
+ - For full model, please see: https://huggingface.co/hfl/llama-3-chinese-8b-instruct-v3
41
+ - If you have questions/issues regarding this model, please submit an issue through https://github.com/ymcui/Chinese-LLaMA-Alpaca-3