louisbrulenaudet commited on
Commit
d2b2009
·
verified ·
1 Parent(s): de9cbce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -6
README.md CHANGED
@@ -17,8 +17,12 @@ tags:
17
  - llama-cpp
18
  - gguf-my-repo
19
  ---
20
-
21
- # louisbrulenaudet/OCRonos-Vintage-Q5_K_M-GGUF
 
 
 
 
22
  This model was converted to GGUF format from [`PleIAs/OCRonos-Vintage`](https://huggingface.co/PleIAs/OCRonos-Vintage) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
23
  Refer to the [original model card](https://huggingface.co/PleIAs/OCRonos-Vintage) for more details on the model.
24
 
@@ -33,12 +37,12 @@ Invoke the llama.cpp server or the CLI.
33
 
34
  ### CLI:
35
  ```bash
36
- llama-cli --hf-repo louisbrulenaudet/OCRonos-Vintage-Q5_K_M-GGUF --hf-file ocronos-vintage-q5_k_m.gguf -p "The meaning to life and the universe is"
37
  ```
38
 
39
  ### Server:
40
  ```bash
41
- llama-server --hf-repo louisbrulenaudet/OCRonos-Vintage-Q5_K_M-GGUF --hf-file ocronos-vintage-q5_k_m.gguf -c 2048
42
  ```
43
 
44
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
@@ -55,9 +59,9 @@ cd llama.cpp && LLAMA_CURL=1 make
55
 
56
  Step 3: Run inference through the main binary.
57
  ```
58
- ./llama-cli --hf-repo louisbrulenaudet/OCRonos-Vintage-Q5_K_M-GGUF --hf-file ocronos-vintage-q5_k_m.gguf -p "The meaning to life and the universe is"
59
  ```
60
  or
61
  ```
62
- ./llama-server --hf-repo louisbrulenaudet/OCRonos-Vintage-Q5_K_M-GGUF --hf-file ocronos-vintage-q5_k_m.gguf -c 2048
63
  ```
 
17
  - llama-cpp
18
  - gguf-my-repo
19
  ---
20
+ ## Dataset Description
21
+ - **Repository:** https://huggingface.co/datasets/louisbrulenaudet/OCRonos-Vintage-GGUF
22
+ - **Leaderboard:** N/A
23
+ - **Point of Contact:** [Louis Brulé Naudet](mailto:louisbrulenaudet@icloud.com)
24
+ -
25
+ # louisbrulenaudet/OCRonos-Vintage-GGUF
26
  This model was converted to GGUF format from [`PleIAs/OCRonos-Vintage`](https://huggingface.co/PleIAs/OCRonos-Vintage) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
27
  Refer to the [original model card](https://huggingface.co/PleIAs/OCRonos-Vintage) for more details on the model.
28
 
 
37
 
38
  ### CLI:
39
  ```bash
40
+ llama-cli --hf-repo louisbrulenaudet/OCRonos-Vintage-GGUF --hf-file ocronos-vintage-q5_k_m.gguf -p "The meaning to life and the universe is"
41
  ```
42
 
43
  ### Server:
44
  ```bash
45
+ llama-server --hf-repo louisbrulenaudet/OCRonos-Vintage-GGUF --hf-file ocronos-vintage-q5_k_m.gguf -c 2048
46
  ```
47
 
48
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
 
59
 
60
  Step 3: Run inference through the main binary.
61
  ```
62
+ ./llama-cli --hf-repo louisbrulenaudet/OCRonos-Vintage-GGUF --hf-file ocronos-vintage-q5_k_m.gguf -p "The meaning to life and the universe is"
63
  ```
64
  or
65
  ```
66
+ ./llama-server --hf-repo louisbrulenaudet/OCRonos-Vintage-GGUF --hf-file ocronos-vintage-q5_k_m.gguf -c 2048
67
  ```