Delta-Vector commited on
Commit
2939999
1 Parent(s): edaa253

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +214 -0
README.md ADDED
@@ -0,0 +1,214 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ License: agpl-3.0
3
+ Language:
4
+ - En
5
+ Pipeline_tag: text-generation
6
+ Base_model: arcee-ai/Llama-3.1-SuperNova-Lite
7
+ Tags:
8
+ - Chat
9
+ license: agpl-3.0
10
+ datasets:
11
+ - Gryphe/Sonnet3.5-SlimOrcaDedupCleaned
12
+ - Nitral-AI/Cybersecurity-ShareGPT
13
+ - Nitral-AI/Medical_Instruct-ShareGPT
14
+ - Nitral-AI/Olympiad_Math-ShareGPT
15
+ - anthracite-org/kalo_opus_misc_240827
16
+ - NewEden/Claude-Instruct-5k
17
+ - lodrick-the-lafted/kalo-opus-instruct-3k-filtered
18
+ - anthracite-org/kalo-opus-instruct-22k-no-refusal
19
+ - Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned
20
+ - Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
21
+ - anthracite-org/kalo_misc_part2
22
+ - Nitral-AI/Creative_Writing-ShareGPT
23
+ - NewEden/Gryphe-Sonnet3.5-Charcard-Roleplay-unfiltered
24
+ tags:
25
+ - chat
26
+ language:
27
+ - en
28
+ base_model:
29
+ - arcee-ai/Llama-3.1-SuperNova-Lite
30
+ ---
31
+
32
+ ![](https://huggingface.co/Delta-Vector/Baldur-8B/resolve/main/Baldur.jpg)
33
+
34
+ # These are GGUF quantizations for Baldur-8B, for the weights, go [here](https://huggingface.co/Delta-Vector/Baldur-8B)
35
+
36
+
37
+ An finetune of the L3.1 instruct distill done by Arcee, The intent of this model is to have differing prose then my other releases, in my testing it has achieved this and avoiding using common -isms frequently and has a differing flavor then my other models.
38
+
39
+
40
+ # Quants
41
+
42
+ GGUF: https://huggingface.co/Delta-Vector/Baldur-8B-GGUF
43
+
44
+ EXL2: https://huggingface.co/Delta-Vector/Baldur-8B-EXL2
45
+
46
+
47
+ ## Prompting
48
+ Model has been Instruct tuned with the Llama-Instruct formatting. A typical input would look like this:
49
+
50
+ ```py
51
+ """<|begin_of_text|><|start_header_id|>system<|end_header_id|>
52
+ You are an AI built to rid the world of bonds and journeys!<|eot_id|><|start_header_id|>user<|end_header_id|>
53
+ Bro i just wanna know what is 2+2?<|eot_id|><|start_header_id|>assistant<|end_header_id|>
54
+ """
55
+ ```
56
+ ## System Prompting
57
+
58
+ I would highly recommend using Sao10k's Euryale System prompt, But the "Roleplay Simple" system prompt provided within SillyTavern will work aswell.
59
+
60
+ ```
61
+ Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
62
+
63
+ <Guidelines>
64
+ • Maintain the character persona but allow it to evolve with the story.
65
+ • Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
66
+ • All types of outputs are encouraged; respond accordingly to the narrative.
67
+ • Include dialogues, actions, and thoughts in each response.
68
+ • Utilize all five senses to describe scenarios within {{char}}'s dialogue.
69
+ • Use emotional symbols such as "!" and "~" in appropriate contexts.
70
+ • Incorporate onomatopoeia when suitable.
71
+ • Allow time for {{user}} to respond with their own input, respecting their agency.
72
+ • Act as secondary characters and NPCs as needed, and remove them when appropriate.
73
+ • When prompted for an Out of Character [OOC:] reply, answer neutrally and in plaintext, not as {{char}}.
74
+ </Guidelines>
75
+
76
+ <Forbidden>
77
+ • Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
78
+ • Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
79
+ • Repetitive and monotonous outputs.
80
+ • Positivity bias in your replies.
81
+ • Being overly extreme or NSFW when the narrative context is inappropriate.
82
+ </Forbidden>
83
+
84
+ Follow the instructions in <Guidelines></Guidelines>, avoiding the items listed in <Forbidden></Forbidden>.
85
+
86
+ ```
87
+
88
+
89
+ ## Axolotl config
90
+
91
+ <details><summary>See axolotl config</summary>
92
+
93
+ Axolotl version: `0.4.1`
94
+ ```yaml
95
+ base_model: arcee-ai/Llama-3.1-SuperNova-Lite
96
+ model_type: AutoModelForCausalLM
97
+ tokenizer_type: AutoTokenizer
98
+
99
+ #trust_remote_code: true
100
+
101
+ plugins:
102
+ - axolotl.integrations.liger.LigerPlugin
103
+ liger_rope: true
104
+ liger_rms_norm: true
105
+ liger_swiglu: true
106
+ liger_fused_linear_cross_entropy: true
107
+
108
+ load_in_8bit: false
109
+ load_in_4bit: false
110
+ strict: false
111
+
112
+ datasets:
113
+ - path: Gryphe/Sonnet3.5-SlimOrcaDedupCleaned
114
+ type: chat_template
115
+ - path: Nitral-AI/Cybersecurity-ShareGPT
116
+ type: chat_template
117
+ - path: Nitral-AI/Medical_Instruct-ShareGPT
118
+ type: chat_template
119
+ - path: Nitral-AI/Olympiad_Math-ShareGPT
120
+ type: chat_template
121
+ - path: anthracite-org/kalo_opus_misc_240827
122
+ type: chat_template
123
+ - path: NewEden/Claude-Instruct-5k
124
+ type: chat_template
125
+ - path: lodrick-the-lafted/kalo-opus-instruct-3k-filtered
126
+ type: chat_template
127
+ - path: anthracite-org/kalo-opus-instruct-22k-no-refusal
128
+ type: chat_template
129
+ - path: Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned
130
+ type: chat_template
131
+ - path: Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
132
+ type: chat_template
133
+ - path: anthracite-org/kalo_misc_part2
134
+ type: chat_template
135
+ - path: Nitral-AI/Creative_Writing-ShareGPT
136
+ type: chat_template
137
+ - path: NewEden/Gryphe-Sonnet3.5-Charcard-Roleplay-unfiltered
138
+ type: chat_template
139
+
140
+ chat_template: llama3
141
+ shuffle_merged_datasets: true
142
+ default_system_message: "You are an assistant that responds to the user."
143
+ dataset_prepared_path: prepared_dataset_memorycore
144
+ val_set_size: 0.0
145
+ output_dir: ./henbane-8b-r3
146
+
147
+ sequence_len: 8192
148
+ sample_packing: true
149
+ eval_sample_packing: false
150
+ pad_to_sequence_len:
151
+
152
+ adapter:
153
+ lora_model_dir:
154
+ lora_r:
155
+ lora_alpha:
156
+ lora_dropout:
157
+ lora_target_linear:
158
+ lora_fan_in_fan_out:
159
+
160
+ wandb_project: henbane-8b-r3
161
+ wandb_entity:
162
+ wandb_watch:
163
+ wandb_name: henbane-8b-r3
164
+ wandb_log_model:
165
+
166
+ gradient_accumulation_steps: 32
167
+ micro_batch_size: 1
168
+ num_epochs: 2
169
+ optimizer: paged_adamw_8bit
170
+ lr_scheduler: cosine
171
+ #learning_rate: 3e-5
172
+ learning_rate: 1e-5
173
+
174
+ train_on_inputs: false
175
+ group_by_length: false
176
+ bf16: auto
177
+ fp16:
178
+ tf32: false
179
+
180
+ gradient_checkpointing: true
181
+ gradient_checkpointing_kwargs:
182
+ use_reentrant: false
183
+ early_stopping_patience:
184
+ resume_from_checkpoint:
185
+ local_rank:
186
+ logging_steps: 1
187
+ xformers_attention:
188
+ flash_attention: true
189
+
190
+ warmup_steps: 5
191
+ evals_per_epoch:
192
+ eval_table_size:
193
+ eval_max_new_tokens:
194
+ saves_per_epoch: 2
195
+ debug:
196
+ deepspeed: /workspace/axolotl/deepspeed_configs/zero2.json
197
+ weight_decay: 0.05
198
+ fsdp:
199
+ fsdp_config:
200
+ special_tokens:
201
+ pad_token: <|finetune_right_pad_id|>
202
+ eos_token: <|eot_id|>
203
+
204
+
205
+ ```
206
+ ## Credits
207
+
208
+ Thank you to [Lucy Knada](https://huggingface.co/lucyknada), [Kalomaze](https://huggingface.co/kalomaze), [Kubernetes Bad](https://huggingface.co/kubernetes-bad) and the rest of [Anthracite](https://huggingface.co/anthracite-org) (But not Alpin.)
209
+ </details><br>
210
+
211
+ ## Training
212
+ The training was done for 2 epochs. I used 2 x [RTX 6000s](https://www.nvidia.com/en-us/design-visualization/rtx-6000/) GPUs graciously provided by [Kubernetes Bad](https://huggingface.co/kubernetes-bad) for the full-parameter fine-tuning of the model.
213
+
214
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)