Update README.md
Browse files
README.md
CHANGED
|
@@ -1,11 +1,28 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3 |
---
|
| 4 |
|
| 5 |
This repository contains the quantized DISC-MedLLM, version of Baichuan-13b-base as the base model.
|
| 6 |
|
| 7 |
The weights are converted to GGML format using [baichuan13b.cpp](https://github.com/ouwei2013/baichuan13b.cpp) (based on [llama.cpp](https://github.com/ggerganov/llama.cpp))
|
| 8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
## How to inference
|
| 10 |
1. [Compile baichuan13b](https://github.com/ouwei2013/baichuan13b.cpp#build), a main executable `baichuan13b/build/bin/main` and a server `baichuan13b/build/bin/server` will be generated.
|
| 11 |
2. Download the weight in this repository to `baichuan13b/build/bin/`
|
|
@@ -41,4 +58,4 @@ llm_output = requests.post(
|
|
| 41 |
"n_predict": 512
|
| 42 |
}).json()
|
| 43 |
print(llm_output)
|
| 44 |
-
```
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
+
datasets:
|
| 4 |
+
- Flmc/DISC-Med-SFT
|
| 5 |
+
language:
|
| 6 |
+
- zh
|
| 7 |
+
pipeline_tag: text-generation
|
| 8 |
+
tags:
|
| 9 |
+
- baichuan
|
| 10 |
+
- medical
|
| 11 |
+
- ggml
|
| 12 |
---
|
| 13 |
|
| 14 |
This repository contains the quantized DISC-MedLLM, version of Baichuan-13b-base as the base model.
|
| 15 |
|
| 16 |
The weights are converted to GGML format using [baichuan13b.cpp](https://github.com/ouwei2013/baichuan13b.cpp) (based on [llama.cpp](https://github.com/ggerganov/llama.cpp))
|
| 17 |
|
| 18 |
+
|Model |GGML quantize method| HDD size |
|
| 19 |
+
|--------------------|--------------------|----------|
|
| 20 |
+
|ggml-model-q4_0.bin | q4_0 | 7.55 GB |
|
| 21 |
+
|ggml-model-q4_1.bin | q4_1 | 8.36 GB |
|
| 22 |
+
|ggml-model-q5_0.bin | q5_0 | 9.17 GB |
|
| 23 |
+
|ggml-model-q5_1.bin | q5_1 | 9.97 GB |
|
| 24 |
+
<!-- |ggml-model-q8_0.bin | q8_0 | ?.?? GB | -->
|
| 25 |
+
|
| 26 |
## How to inference
|
| 27 |
1. [Compile baichuan13b](https://github.com/ouwei2013/baichuan13b.cpp#build), a main executable `baichuan13b/build/bin/main` and a server `baichuan13b/build/bin/server` will be generated.
|
| 28 |
2. Download the weight in this repository to `baichuan13b/build/bin/`
|
|
|
|
| 58 |
"n_predict": 512
|
| 59 |
}).json()
|
| 60 |
print(llm_output)
|
| 61 |
+
```
|