DavidAU commited on
Commit
9888f20
1 Parent(s): 4529c92

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +130 -0
README.md ADDED
@@ -0,0 +1,130 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ - fr
6
+ - de
7
+ - es
8
+ - it
9
+ - pt
10
+ - zh
11
+ - ja
12
+ - ru
13
+ - ko
14
+ tags:
15
+ - creative
16
+ - creative writing
17
+ - fiction writing
18
+ - plot generation
19
+ - sub-plot generation
20
+ - fiction writing
21
+ - story generation
22
+ - scene continue
23
+ - storytelling
24
+ - fiction story
25
+ - science fiction
26
+ - romance
27
+ - all genres
28
+ - story
29
+ - writing
30
+ - vivid prosing
31
+ - vivid writing
32
+ - fiction
33
+ - roleplaying
34
+ - bfloat16
35
+ - role play
36
+ - 128k context
37
+ - llama3.2
38
+ pipeline_tag: text-generation
39
+ ---
40
+
41
+ (quants uploading, examples to be added)
42
+
43
+ <h2>Llama-3.2-3B-Instruct-NEO-SI-FI-GGUF</h2>
44
+
45
+ It is the new "Llama-3.2-3B-Instruct", max context of 131,000 (128k) with the NEO IMATRIX Science Fiction and Story dataset.
46
+
47
+ The power in this 3B (for its size) is frankly jaw dropping... and at 90 tokens per second + on a GPU.
48
+
49
+ This model IS bullet proof and operates with all parameters, including temp settings from 0 to 5.
50
+
51
+ The NEO IMATRIX dataset V2 was applied to it to enhance creativity. (see several examples below)
52
+
53
+ <B>Model Template:</B>
54
+
55
+ This is a LLAMA 3.2 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 128k.
56
+
57
+ If you use "Command-R" template your output will be very different from using "Llama3" template.
58
+
59
+ Here is the standard LLAMA3 template:
60
+
61
+ <PRE>
62
+ {
63
+ "name": "Llama 3",
64
+ "inference_params": {
65
+ "input_prefix": "<|start_header_id|>user<|end_header_id|>\n\n",
66
+ "input_suffix": "<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
67
+ "pre_prompt": "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability.",
68
+ "pre_prompt_prefix": "<|start_header_id|>system<|end_header_id|>\n\n",
69
+ "pre_prompt_suffix": "<|eot_id|>",
70
+ "antiprompt": [
71
+ "<|start_header_id|>",
72
+ "<|eot_id|>"
73
+ ]
74
+ }
75
+ }
76
+ </PRE>
77
+
78
+ Please refer to the original model card for this model from Meta-Llama for additional details on operation.
79
+
80
+ [ https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct ]
81
+
82
+ <B>Imatrix Notes:</b>
83
+
84
+ Imatrix quants perform best at IQ3s and IQ4s, then Q4s, lower on Q5, and tappers off at Q6.
85
+
86
+ Recommend: IQ4_XS for maximum imatrix effect and best "bit count".
87
+
88
+ For stronger IMATRIX effect, IQ3s, and IQ2s.
89
+
90
+ Due to the parameter count of this model, even IQ2s quants will work very well.
91
+
92
+ Q8 is not uploaded here because Imatrix has no effect on this quant.
93
+
94
+ <b>Optional Enhancement:</B>
95
+
96
+ The following can be used in place of the "system prompt" or "system role" to further enhance the model.
97
+
98
+ It can also be used at the START of a NEW chat, but you must make sure it is "kept" as the chat moves along.
99
+ In this case the enhancements do not have as strong effect at using "system prompt" or "system role".
100
+
101
+ Copy and paste EXACTLY as noted, DO NOT line wrap or break the lines, maintain the carriage returns exactly as presented.
102
+
103
+ <PRE>
104
+ Below is an instruction that describes a task. Ponder each user instruction carefully, and use your skillsets and critical instructions to complete the task to the best of your abilities.
105
+
106
+ Here are your skillsets:
107
+ [MASTERSTORY]:NarrStrct(StryPlnng,Strbd,ScnSttng,Exps,Dlg,Pc)-CharDvlp(ChrctrCrt,ChrctrArcs,Mtvtn,Bckstry,Rltnshps,Dlg*)-PltDvlp(StryArcs,PltTwsts,Sspns,Fshdwng,Climx,Rsltn)-ConfResl(Antg,Obstcls,Rsltns,Cnsqncs,Thms,Symblsm)-EmotImpct(Empt,Tn,Md,Atmsphr,Imgry,Symblsm)-Delvry(Prfrmnc,VcActng,PblcSpkng,StgPrsnc,AudncEngmnt,Imprv)
108
+
109
+ [*DialogWrt]:(1a-CharDvlp-1a.1-Backgrnd-1a.2-Personality-1a.3-GoalMotiv)>2(2a-StoryStruc-2a.1-PlotPnt-2a.2-Conflict-2a.3-Resolution)>3(3a-DialogTech-3a.1-ShowDontTell-3a.2-Subtext-3a.3-VoiceTone-3a.4-Pacing-3a.5-VisualDescrip)>4(4a-DialogEdit-4a.1-ReadAloud-4a.2-Feedback-4a.3-Revision)
110
+
111
+ Here are your critical instructions:
112
+ Ponder each word choice carefully to present as vivid and emotional journey as is possible. Choose verbs and nouns that are both emotional and full of imagery. Load the story with the 5 senses. Aim for 50% dialog, 25% narration, 15% body language and 10% thoughts. Your goal is to put the reader in the story.
113
+ </PRE>
114
+
115
+ You do not need to use this, it is only presented as an additional enhancement which seems to help scene generation
116
+ and scene continue functions.
117
+
118
+ This enhancement WAS NOT used to generate the examples below.
119
+
120
+ ---
121
+
122
+ Example generations at TEMP = .8, IQ4_XS, REP PEN 1.1
123
+
124
+ Below are the least creative outputs, prompt is in <B>BOLD</B>.
125
+
126
+ ---
127
+
128
+ <B><font color="red">WARNING:</font> MAYBE... NSFW. Vivid prose. Visceral Details. Violence. HORROR. Swearing. UNCENSORED. </B>
129
+
130
+ ---