Update README.md
Browse files
README.md
CHANGED
|
@@ -14,6 +14,38 @@ base_model: DavidAU/L3-MOE-4X8B-Grand-Horror-25B
|
|
| 14 |
This model was converted to GGUF format from [`DavidAU/L3-MOE-4X8B-Grand-Horror-25B`](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
| 15 |
Refer to the [original model card](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) for more details on the model.
|
| 16 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
## Use with llama.cpp
|
| 18 |
Install llama.cpp through brew (works on Mac and Linux)
|
| 19 |
|
|
|
|
| 14 |
This model was converted to GGUF format from [`DavidAU/L3-MOE-4X8B-Grand-Horror-25B`](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
| 15 |
Refer to the [original model card](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) for more details on the model.
|
| 16 |
|
| 17 |
+
---
|
| 18 |
+
It is a LLama3 model, max context of 8192 (or 32k+ with rope) using mixture of experts to combine Dark/Horror models models of 8B each into one massive powerhouse at 25B parameters (equal to 32B - 4 X 8 B).
|
| 19 |
+
|
| 20 |
+
This model's instruction following, and output generation for creative writing, prose, fiction and role play are exceptional.
|
| 21 |
+
|
| 22 |
+
It excels at description, dialog, imagery, metaphors, and prose - and shows great variations in sentence / paragraph size, length, and composition.
|
| 23 |
+
|
| 24 |
+
It is also not afraid, and will not pull its punches.
|
| 25 |
+
|
| 26 |
+
And it has a sense of humor too.
|
| 27 |
+
|
| 28 |
+
It can do horror just as easily as it can do romance.
|
| 29 |
+
|
| 30 |
+
Most notably dialog is very "un-ai" like, combined with prose (short, and terse at times).
|
| 31 |
+
|
| 32 |
+
(lots of different examples below, including 2, 3 and 4 experts and different genres)
|
| 33 |
+
|
| 34 |
+
And it is fast: 34 t/s (2 experts) on a low end 16GB card, Q3KS.
|
| 35 |
+
|
| 36 |
+
Double this speed for standard/mid-range video cards.
|
| 37 |
+
|
| 38 |
+
Model can be used also for all genres (examples below showing this).
|
| 39 |
+
|
| 40 |
+
This model has been designed to be relatively bullet proof and operates with all parameters, including temp settings from 0 to 5.
|
| 41 |
+
|
| 42 |
+
It is an extraordinary compressed model, with a very low perplexity level (lower than Meta Llama3 Instruct).
|
| 43 |
+
|
| 44 |
+
It is for any writing, fiction or roleplay activity.
|
| 45 |
+
|
| 46 |
+
It requires Llama3 template and/or "Command-R" template.
|
| 47 |
+
|
| 48 |
+
---
|
| 49 |
## Use with llama.cpp
|
| 50 |
Install llama.cpp through brew (works on Mac and Linux)
|
| 51 |
|