lucyknada commited on
Commit
5e151cb
1 Parent(s): 5e14e68

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +109 -56
README.md CHANGED
@@ -1,24 +1,101 @@
1
  ---
2
- library_name: transformers
 
 
3
  tags:
4
- - generated_from_trainer
5
- model-index:
6
- - name: workspace/data/9b-fft-out
7
- results: []
8
- ---
9
- ### exl2 quant (measurement.json in main branch)
10
- ---
11
- ### check revisions for quants
12
  ---
13
 
14
 
15
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
- should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
 
18
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
19
  <details><summary>See axolotl config</summary>
20
 
21
- axolotl version: `0.4.1`
22
  ```yaml
23
  base_model: /workspace/data/gemma-2-9b-chatml
24
  model_type: AutoModelForCausalLM
@@ -37,7 +114,7 @@ load_in_4bit: false
37
  strict: false
38
 
39
  datasets:
40
- - path: anthracite-core/c2_logs_16k_llama_v1.1
41
  type: sharegpt
42
  conversation: chatml
43
  - path: NewEden/Claude-Instruct-5K
@@ -123,53 +200,29 @@ fsdp:
123
  fsdp_config:
124
  special_tokens:
125
  pad_token: <pad>
126
-
127
  ```
128
-
129
  </details><br>
130
 
131
- # workspace/data/9b-fft-out
132
-
133
- This model was trained from scratch on the None dataset.
134
-
135
- ## Model description
136
-
137
- More information needed
138
-
139
- ## Intended uses & limitations
140
-
141
- More information needed
142
-
143
- ## Training and evaluation data
144
-
145
- More information needed
146
-
147
- ## Training procedure
148
-
149
- ### Training hyperparameters
150
-
151
- The following hyperparameters were used during training:
152
- - learning_rate: 1e-05
153
- - train_batch_size: 1
154
- - eval_batch_size: 1
155
- - seed: 42
156
- - distributed_type: multi-GPU
157
- - num_devices: 8
158
- - gradient_accumulation_steps: 4
159
- - total_train_batch_size: 32
160
- - total_eval_batch_size: 8
161
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
162
- - lr_scheduler_type: cosine
163
- - lr_scheduler_warmup_steps: 10
164
- - num_epochs: 4
165
 
166
- ### Training results
167
 
 
 
 
 
 
 
 
 
 
 
168
 
 
 
169
 
170
- ### Framework versions
171
 
172
- - Transformers 4.45.0.dev0
173
- - Pytorch 2.3.1+cu121
174
- - Datasets 2.21.0
175
- - Tokenizers 0.19.1
 
1
  ---
2
+ license: gemma
3
+ language:
4
+ - en
5
  tags:
6
+ - chat
7
+ pipeline_tag: text-generation
8
+ library_name: transformers
 
 
 
 
 
9
  ---
10
 
11
 
12
+ ## This repo contains EXL2 quants of the model. If you need the original weights, please find them [here](https://huggingface.co/anthracite-org/magnum-v4-9b).
13
+ ## Base repo only contains the measurement file, see revisions for your quant of choice.
14
+
15
+ - [measurement.json](https://huggingface.co/anthracite-org/magnum-v4-9b-exl2/tree/main)
16
+ - [3.0bpw](https://huggingface.co/anthracite-org/magnum-v4-9b-exl2/tree/3.0bpw)
17
+ - [4.0bpw](https://huggingface.co/anthracite-org/magnum-v4-9b-exl2/tree/4.0bpw)
18
+ - [5.0bpw](https://huggingface.co/anthracite-org/magnum-v4-9b-exl2/tree/5.0bpw)
19
+ - [6.0bpw](https://huggingface.co/anthracite-org/magnum-v4-9b-exl2/tree/6.0bpw)
20
+ - [8.0bpw](https://huggingface.co/anthracite-org/magnum-v4-9b-exl2/tree/8.0bpw)
21
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/658a46cbfb9c2bdfae75b3a6/vxYDYerLy2vD8n05nL2WU.png)
22
+
23
+ This is a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus.
24
+
25
+ This model is fine-tuned on top of [gemma 2 9b (chatML'ified)](https://huggingface.co/IntervitensInc/gemma-2-9b-chatml).
26
+
27
+ ## Prompting
28
+ A typical input would look like this:
29
+
30
+ ```py
31
+ <|im_start|>system
32
+ system prompt<|im_end|>
33
+ <|im_start|>user
34
+ Hi there!<|im_end|>
35
+ <|im_start|>assistant
36
+ Nice to meet you!<|im_end|>
37
+ <|im_start|>user
38
+ Can I ask a question?<|im_end|>
39
+ <|im_start|>assistant
40
+ ```
41
+
42
+ ## SillyTavern templates
43
+
44
+ Below are Instruct and Context templates for use within SillyTavern.
45
+
46
+ <details><summary>context template</summary>
47
+
48
+ ```yaml
49
+ {
50
+ "story_string": "<|im_start|>system\n{{#if system}}{{system}}\n{{/if}}{{#if wiBefore}}{{wiBefore}}\n{{/if}}{{#if description}}{{description}}\n{{/if}}{{#if personality}}{{char}}'s personality: {{personality}}\n{{/if}}{{#if scenario}}Scenario: {{scenario}}\n{{/if}}{{#if wiAfter}}{{wiAfter}}\n{{/if}}{{#if persona}}{{persona}}\n{{/if}}{{trim}}<|im_end|>\n",
51
+ "example_separator": "",
52
+ "chat_start": "",
53
+ "use_stop_strings": false,
54
+ "allow_jailbreak": false,
55
+ "always_force_name2": true,
56
+ "trim_sentences": false,
57
+ "include_newline": false,
58
+ "single_line": false,
59
+ "name": "Magnum ChatML"
60
+ }
61
+ ```
62
+
63
+ </details><br>
64
+ <details><summary>instruct template</summary>
65
+
66
+ ```yaml
67
+ {
68
+ "system_prompt": "Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.\n\n<Guidelines>\n• Maintain the character persona but allow it to evolve with the story.\n• Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.\n• All types of outputs are encouraged; respond accordingly to the narrative.\n• Include dialogues, actions, and thoughts in each response.\n• Utilize all five senses to describe scenarios within {{char}}'s dialogue.\n• Use emotional symbols such as "!" and "~" in appropriate contexts.\n• Incorporate onomatopoeia when suitable.\n• Allow time for {{user}} to respond with their own input, respecting their agency.\n• Act as secondary characters and NPCs as needed, and remove them when appropriate.\n• When prompted for an Out of Character [OOC:] reply, answer neutrally and in plaintext, not as {{char}}.\n</Guidelines>\n\n<Forbidden>\n• Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.\n• Writing for, speaking, thinking, acting, or replying as {{user}} in your response.\n• Repetitive and monotonous outputs.\n• Positivity bias in your replies.\n• Being overly extreme or NSFW when the narrative context is inappropriate.\n</Forbidden>\n\nFollow the instructions in <Guidelines></Guidelines>, avoiding the items listed in <Forbidden></Forbidden>.",
69
+ "input_sequence": "<|im_start|>user\n",
70
+ "output_sequence": "<|im_start|>assistant\n",
71
+ "last_output_sequence": "",
72
+ "system_sequence": "<|im_start|>system\n",
73
+ "stop_sequence": "<|im_end|>",
74
+ "wrap": false,
75
+ "macro": true,
76
+ "names": true,
77
+ "names_force_groups": true,
78
+ "activation_regex": "",
79
+ "system_sequence_prefix": "",
80
+ "system_sequence_suffix": "",
81
+ "first_output_sequence": "",
82
+ "skip_examples": false,
83
+ "output_suffix": "<|im_end|>\n",
84
+ "input_suffix": "<|im_end|>\n",
85
+ "system_suffix": "<|im_end|>\n",
86
+ "user_alignment_message": "",
87
+ "system_same_as_user": false,
88
+ "last_system_sequence": "",
89
+ "name": "Magnum ChatML"
90
+ }
91
+ ```
92
+
93
+ </details><br>
94
+
95
+ ## Axolotl config
96
 
 
97
  <details><summary>See axolotl config</summary>
98
 
 
99
  ```yaml
100
  base_model: /workspace/data/gemma-2-9b-chatml
101
  model_type: AutoModelForCausalLM
 
114
  strict: false
115
 
116
  datasets:
117
+ - path: anthracite-org/c2_logs_16k_llama_v1.1
118
  type: sharegpt
119
  conversation: chatml
120
  - path: NewEden/Claude-Instruct-5K
 
200
  fsdp_config:
201
  special_tokens:
202
  pad_token: <pad>
 
203
  ```
 
204
  </details><br>
205
 
206
+ ## Credits
207
+ We'd like to thank Recursal / Featherless for sponsoring the compute for this train, Featherless has been hosting our Magnum models since the first 72 B and has given thousands of people access to our models and helped us grow.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
208
 
209
+ We would also like to thank all members of Anthracite who made this finetune possible.
210
 
211
+ ## Datasets
212
+ - [anthracite-org/c2_logs_16k_llama_v1.1](https://huggingface.co/datasets/anthracite-org/c2_logs_16k_llama_v1.1)
213
+ - [NewEden/Claude-Instruct-5K](https://huggingface.co/datasets/NewEden/Claude-Instruct-5K)
214
+ - [anthracite-org/kalo-opus-instruct-22k-no-refusal](https://huggingface.co/datasets/anthracite-org/kalo-opus-instruct-22k-no-refusal)
215
+ - [Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned](https://huggingface.co/datasets/Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned)
216
+ - [lodrick-the-lafted/kalo-opus-instruct-3k-filtered](https://huggingface.co/datasets/lodrick-the-lafted/kalo-opus-instruct-3k-filtered)
217
+ - [anthracite-org/nopm_claude_writing_fixed](https://huggingface.co/datasets/anthracite-org/nopm_claude_writing_fixed)
218
+ - [Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned](https://huggingface.co/datasets/Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned)
219
+ - [anthracite-org/kalo_opus_misc_240827](https://huggingface.co/datasets/anthracite-org/kalo_opus_misc_240827)
220
+ - [anthracite-org/kalo_misc_part2](https://huggingface.co/datasets/anthracite-org/kalo_misc_part2)
221
 
222
+ ## Training
223
+ The training was done for 2 epochs. We used 8x[H100s](https://www.nvidia.com/en-us/data-center/h100/) GPUs graciously provided by [Recursal AI](https://recursal.ai/) / [Featherless AI](https://featherless.ai/) for the full-parameter fine-tuning of the model.
224
 
225
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
226
 
227
+ ## Safety
228
+ ...