itlwas commited on
Commit
2f698d5
·
verified ·
1 Parent(s): 7e82cf2

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +10 -25
README.md CHANGED
@@ -13,49 +13,34 @@ widget:
13
  content: What is your favorite condiment?
14
  extra_gated_description: If you want to learn more about how we process your personal
15
  data, please read our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
16
- base_model: mistralai/Mistral-7B-Instruct-v0.2
17
  ---
18
 
19
  # AIronMind/Mistral-7B-Instruct-v0.2-Q4_K_M-GGUF
20
  This model was converted to GGUF format from [`mistralai/Mistral-7B-Instruct-v0.2`](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
21
  Refer to the [original model card](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) for more details on the model.
22
-
23
  ## Use with llama.cpp
24
- Install llama.cpp through brew (works on Mac and Linux)
25
 
26
- ```bash
27
- brew install llama.cpp
28
 
 
 
29
  ```
30
  Invoke the llama.cpp server or the CLI.
31
 
32
- ### CLI:
33
- ```bash
34
- llama-cli --hf-repo AIronMind/Mistral-7B-Instruct-v0.2-Q4_K_M-GGUF --hf-file mistral-7b-instruct-v0.2-q4_k_m.gguf -p "The meaning to life and the universe is"
35
- ```
36
 
37
- ### Server:
38
  ```bash
39
- llama-server --hf-repo AIronMind/Mistral-7B-Instruct-v0.2-Q4_K_M-GGUF --hf-file mistral-7b-instruct-v0.2-q4_k_m.gguf -c 2048
40
  ```
41
 
42
- Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
43
 
44
- Step 1: Clone llama.cpp from GitHub.
45
- ```
46
- git clone https://github.com/ggerganov/llama.cpp
47
  ```
48
 
49
- Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
50
- ```
51
- cd llama.cpp && LLAMA_CURL=1 make
52
- ```
53
 
54
- Step 3: Run inference through the main binary.
55
- ```
56
- ./llama-cli --hf-repo AIronMind/Mistral-7B-Instruct-v0.2-Q4_K_M-GGUF --hf-file mistral-7b-instruct-v0.2-q4_k_m.gguf -p "The meaning to life and the universe is"
57
- ```
58
- or
59
  ```
60
- ./llama-server --hf-repo AIronMind/Mistral-7B-Instruct-v0.2-Q4_K_M-GGUF --hf-file mistral-7b-instruct-v0.2-q4_k_m.gguf -c 2048
61
  ```
 
13
  content: What is your favorite condiment?
14
  extra_gated_description: If you want to learn more about how we process your personal
15
  data, please read our <a href="https://mistral.ai/terms/">Privacy Policy</a>.
 
16
  ---
17
 
18
  # AIronMind/Mistral-7B-Instruct-v0.2-Q4_K_M-GGUF
19
  This model was converted to GGUF format from [`mistralai/Mistral-7B-Instruct-v0.2`](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
20
  Refer to the [original model card](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) for more details on the model.
 
21
  ## Use with llama.cpp
 
22
 
23
+ Install llama.cpp through brew.
 
24
 
25
+ ```bash
26
+ brew install ggerganov/ggerganov/llama.cpp
27
  ```
28
  Invoke the llama.cpp server or the CLI.
29
 
30
+ CLI:
 
 
 
31
 
 
32
  ```bash
33
+ llama-cli --hf-repo AIronMind/Mistral-7B-Instruct-v0.2-Q4_K_M-GGUF --model mistral-7b-instruct-v0.2.Q4_K_M.gguf -p "The meaning to life and the universe is"
34
  ```
35
 
36
+ Server:
37
 
38
+ ```bash
39
+ llama-server --hf-repo AIronMind/Mistral-7B-Instruct-v0.2-Q4_K_M-GGUF --model mistral-7b-instruct-v0.2.Q4_K_M.gguf -c 2048
 
40
  ```
41
 
42
+ Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
 
 
 
43
 
 
 
 
 
 
44
  ```
45
+ git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m mistral-7b-instruct-v0.2.Q4_K_M.gguf -n 128
46
  ```