hansmueller464 commited on
Commit
f44eca1
1 Parent(s): 269d8a1

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +58 -0
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: cc-by-nc-4.0
5
+ library_name: transformers
6
+ tags:
7
+ - biology
8
+ - medical
9
+ - llama-cpp
10
+ - gguf-my-repo
11
+ datasets:
12
+ - argilla/dpo-mix-7k
13
+ - nvidia/HelpSteer
14
+ - jondurbin/airoboros-3.2
15
+ - hkust-nlp/deita-10k-v0
16
+ - LDJnr/Capybara
17
+ - HPAI-BSC/CareQA
18
+ - GBaker/MedQA-USMLE-4-options
19
+ - lukaemon/mmlu
20
+ - bigbio/pubmed_qa
21
+ - openlifescienceai/medmcqa
22
+ - bigbio/med_qa
23
+ - HPAI-BSC/better-safe-than-sorry
24
+ - HPAI-BSC/pubmedqa-cot
25
+ - HPAI-BSC/medmcqa-cot
26
+ - HPAI-BSC/medqa-cot
27
+ pipeline_tag: question-answering
28
+ ---
29
+
30
+ # hansmueller464/Llama3-Aloe-8B-Alpha-Q6_K-GGUF
31
+ This model was converted to GGUF format from [`HPAI-BSC/Llama3-Aloe-8B-Alpha`](https://huggingface.co/HPAI-BSC/Llama3-Aloe-8B-Alpha) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
32
+ Refer to the [original model card](https://huggingface.co/HPAI-BSC/Llama3-Aloe-8B-Alpha) for more details on the model.
33
+ ## Use with llama.cpp
34
+
35
+ Install llama.cpp through brew.
36
+
37
+ ```bash
38
+ brew install ggerganov/ggerganov/llama.cpp
39
+ ```
40
+ Invoke the llama.cpp server or the CLI.
41
+
42
+ CLI:
43
+
44
+ ```bash
45
+ llama-cli --hf-repo hansmueller464/Llama3-Aloe-8B-Alpha-Q6_K-GGUF --model llama3-aloe-8b-alpha.Q6_K.gguf -p "The meaning to life and the universe is"
46
+ ```
47
+
48
+ Server:
49
+
50
+ ```bash
51
+ llama-server --hf-repo hansmueller464/Llama3-Aloe-8B-Alpha-Q6_K-GGUF --model llama3-aloe-8b-alpha.Q6_K.gguf -c 2048
52
+ ```
53
+
54
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
55
+
56
+ ```
57
+ git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m llama3-aloe-8b-alpha.Q6_K.gguf -n 128
58
+ ```