Triangle104 commited on
Commit
1b831ee
·
verified ·
1 Parent(s): f517277

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -0
README.md CHANGED
@@ -14,6 +14,38 @@ base_model: DavidAU/L3-MOE-4X8B-Grand-Horror-25B
14
  This model was converted to GGUF format from [`DavidAU/L3-MOE-4X8B-Grand-Horror-25B`](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
15
  Refer to the [original model card](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) for more details on the model.
16
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  ## Use with llama.cpp
18
  Install llama.cpp through brew (works on Mac and Linux)
19
 
 
14
  This model was converted to GGUF format from [`DavidAU/L3-MOE-4X8B-Grand-Horror-25B`](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
15
  Refer to the [original model card](https://huggingface.co/DavidAU/L3-MOE-4X8B-Grand-Horror-25B) for more details on the model.
16
 
17
+ ---
18
+ It is a LLama3 model, max context of 8192 (or 32k+ with rope) using mixture of experts to combine Dark/Horror models models of 8B each into one massive powerhouse at 25B parameters (equal to 32B - 4 X 8 B).
19
+
20
+ This model's instruction following, and output generation for creative writing, prose, fiction and role play are exceptional.
21
+
22
+ It excels at description, dialog, imagery, metaphors, and prose - and shows great variations in sentence / paragraph size, length, and composition.
23
+
24
+ It is also not afraid, and will not pull its punches.
25
+
26
+ And it has a sense of humor too.
27
+
28
+ It can do horror just as easily as it can do romance.
29
+
30
+ Most notably dialog is very "un-ai" like, combined with prose (short, and terse at times).
31
+
32
+ (lots of different examples below, including 2, 3 and 4 experts and different genres)
33
+
34
+ And it is fast: 34 t/s (2 experts) on a low end 16GB card, Q3KS.
35
+
36
+ Double this speed for standard/mid-range video cards.
37
+
38
+ Model can be used also for all genres (examples below showing this).
39
+
40
+ This model has been designed to be relatively bullet proof and operates with all parameters, including temp settings from 0 to 5.
41
+
42
+ It is an extraordinary compressed model, with a very low perplexity level (lower than Meta Llama3 Instruct).
43
+
44
+ It is for any writing, fiction or roleplay activity.
45
+
46
+ It requires Llama3 template and/or "Command-R" template.
47
+
48
+ ---
49
  ## Use with llama.cpp
50
  Install llama.cpp through brew (works on Mac and Linux)
51