BlouseJury commited on
Commit
bf54ebb
1 Parent(s): 88f07c2

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +188 -0
README.md ADDED
@@ -0,0 +1,188 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ base_model: meta-llama/Meta-Llama-3-8B
4
+ tags:
5
+ - generated_from_trainer
6
+ - axolotl
7
+ model-index:
8
+ - name: out
9
+ results: []
10
+ datasets:
11
+ - cognitivecomputations/Dolphin-2.9
12
+ - teknium/OpenHermes-2.5
13
+ - m-a-p/CodeFeedback-Filtered-Instruction
14
+ - cognitivecomputations/dolphin-coder
15
+ - cognitivecomputations/samantha-data
16
+ - microsoft/orca-math-word-problems-200k
17
+ - Locutusque/function-calling-chatml
18
+ - internlm/Agent-FLAN
19
+ ---
20
+
21
+ # Dolphin 2.9.1 Llama 3 8b 🐬
22
+
23
+ Curated and trained by Eric Hartford, Lucas Atkins, and Fernando Fernandes, and Cognitive Computations
24
+
25
+ [![Discord](https://img.shields.io/discord/1156064224225808488?logo=Discord&logoColor=%23ffffff&label=Discord&link=https%3A%2F%2Fdiscord.gg%2FtCMkMDDHwm)](https://discord.gg/cognitivecomputations)
26
+ Discord: https://discord.gg/cognitivecomputations
27
+
28
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/63111b2d88942700629f5771/ldkN1J0WIDQwU4vutGYiD.png" width="600" />
29
+
30
+ We have retrained our LLama-3-8b fine tune to address behavioral issues in the initial 2.9 dataset. Specifically, Systemchat was causing the model to be *too* reliant on the system prompt. Additionally, it had an occasional quirk that would cause the model to overly reference the system prompt. We also found generation length was at times not sufficient for any given task. We identified the culprit as Ultrachat. Accounting for these concerns, we removed systemchat and ultrachat from the dataset. It is otherwise identical to dolphin-2.9.
31
+
32
+ Our appreciation for the sponsors of Dolphin 2.9.1:
33
+ - [Crusoe Cloud](https://crusoe.ai/) - provided excellent on-demand 8xL40S node
34
+
35
+ This model is based on Llama-3-8b, and is governed by [META LLAMA 3 COMMUNITY LICENSE AGREEMENT](LICENSE)
36
+
37
+ The base model has 8k context, and the full-weight fine-tuning was with 4k sequence length.
38
+
39
+ It took 1.5 days on an 8x L40S provided by Crusoe Cloud
40
+
41
+ This model was trained FFT on all parameters, using ChatML prompt template format.
42
+
43
+ example:
44
+
45
+ ```
46
+ <|im_start|>system
47
+ You are Dolphin, a helpful AI assistant.<|im_end|>
48
+ <|im_start|>user
49
+ {prompt}<|im_end|>
50
+ <|im_start|>assistant
51
+
52
+ ```
53
+
54
+ Dolphin-2.9.1 has a variety of instruction, conversational, and coding skills. It also has initial agentic abilities and supports function calling.
55
+
56
+ Dolphin is uncensored. We have filtered the dataset to remove alignment and bias. This makes the model more compliant. You are advised to implement your own alignment layer before exposing the model as a service. It will be highly compliant with any requests, even unethical ones. Please read my blog post about uncensored models. https://erichartford.com/uncensored-models You are responsible for any content you create using this model. Enjoy responsibly.
57
+
58
+ Dolphin is licensed according to Meta's Llama license. We grant permission for any use, including commercial, that falls within accordance with Meta's Llama-3 license. Dolphin was trained on data generated from GPT4, among other models.
59
+
60
+ ## Evals
61
+
62
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63111b2d88942700629f5771/0pqSc8jsJlhBH8dcgpwE7.png)
63
+
64
+ ## Training
65
+
66
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
67
+ <details><summary>See axolotl config</summary>
68
+
69
+ axolotl version: `0.4.0`
70
+ ```yaml
71
+ base_model: meta-llama/Meta-Llama-3-8B
72
+ model_type: AutoModelForCausalLM
73
+ tokenizer_type: AutoTokenizer
74
+ tokenizer_use_fast: false
75
+
76
+
77
+ load_in_8bit: false
78
+ load_in_4bit: false
79
+ strict: false
80
+ model_config:
81
+
82
+ datasets:
83
+ - path: /workspace/datasets/dolphin-2.9/dolphin201-sharegpt2.jsonl
84
+ type: sharegpt
85
+ conversation: chatml
86
+ - path: /workspace/datasets/dolphin-2.9/dolphin-coder-translate-sharegpt2.jsonl
87
+ type: sharegpt
88
+ conversation: chatml
89
+ - path: /workspace/datasets/dolphin-2.9/dolphin-coder-codegen-sharegpt2.jsonl
90
+ type: sharegpt
91
+ conversation: chatml
92
+ - path: /workspace/datasets/dolphin-2.9/m-a-p_Code-Feedback-sharegpt-unfiltered.jsonl
93
+ type: sharegpt
94
+ conversation: chatml
95
+ - path: /workspace/datasets/dolphin-2.9/m-a-p_CodeFeedback-Filtered-Instruction-sharegpt-unfiltered.jsonl
96
+ type: sharegpt
97
+ conversation: chatml
98
+ - path: /workspace/datasets/dolphin-2.9/not_samantha_norefusals.jsonl
99
+ type: sharegpt
100
+ conversation: chatml
101
+ - path: /workspace/datasets/dolphin-2.9/Orca-Math-resort-unfiltered.jsonl
102
+ type: sharegpt
103
+ conversation: chatml
104
+ - path: /workspace/datasets/dolphin-2.9/agent_instruct_react_unfiltered.jsonl
105
+ type: sharegpt
106
+ conversation: chatml
107
+ - path: /workspace/datasets/dolphin-2.9/toolbench_instruct_j1s1_3k_unfiltered.jsonl
108
+ type: sharegpt
109
+ conversation: chatml
110
+ - path: /workspace/datasets/dolphin-2.9/toolbench_negative_unfiltered.jsonl
111
+ type: sharegpt
112
+ conversation: chatml
113
+ - path: /workspace/datasets/dolphin-2.9/toolbench_react_10p_unfiltered.jsonl
114
+ type: sharegpt
115
+ conversation: chatml
116
+ - path: /workspace/datasets/dolphin-2.9/toolbench_tflan_cot_30p_unfiltered.jsonl
117
+ type: sharegpt
118
+ conversation: chatml
119
+ - path: /workspace/datasets/dolphin-2.9/openhermes200k_unfiltered.jsonl
120
+ type: sharegpt
121
+ conversation: chatml
122
+
123
+ chat_template: chatml
124
+
125
+
126
+ dataset_prepared_path: /workspace/datasets/dolphin-2.9/thingy
127
+ val_set_size: 0.0002
128
+ output_dir: ./out
129
+
130
+ sequence_len: 4096
131
+ sample_packing: true
132
+ pad_to_sequence_len: true
133
+
134
+ gradient_accumulation_steps: 4
135
+ micro_batch_size: 3
136
+ num_epochs: 3
137
+ logging_steps: 1
138
+ optimizer: adamw_8bit
139
+ lr_scheduler: cosine
140
+ learning_rate: 2e-5
141
+
142
+ wandb_project: dolphin-2.9-mixtral-8x22b
143
+ wandb_watch:
144
+ wandb_run_id:
145
+ wandb_log_model:
146
+
147
+ train_on_inputs: false
148
+ group_by_length: false
149
+ bf16: auto
150
+ fp16:
151
+ tf32: false
152
+
153
+ gradient_checkpointing: true
154
+ gradient_checkpointing_kwargs:
155
+ use_reentrant: false
156
+ early_stopping_patience:
157
+ resume_from_checkpoint:
158
+ local_rank:
159
+ logging_steps: 1
160
+ xformers_attention:
161
+ flash_attention: true
162
+ saves_per_epoch: 4
163
+ save_total_limit: 2
164
+ save_steps:
165
+ evals_per_epoch: 4
166
+ eval_sample_packing: false
167
+ debug:
168
+ deepspeed: deepspeed_configs/zero3_bf16.json
169
+ weight_decay: 0.05
170
+ fsdp:
171
+ fsdp_config:
172
+ special_tokens:
173
+ eos_token: "<|im_end|>"
174
+ pad_token: "<|end_of_text|>"
175
+ tokens:
176
+ - "<|im_start|>"
177
+ - "<|im_end|>"
178
+
179
+ ```
180
+
181
+ </details><br>
182
+
183
+ ### Framework versions
184
+
185
+ - Transformers 4.40.0
186
+ - Pytorch 2.2.2+cu121
187
+ - Datasets 2.18.0
188
+ - Tokenizers 0.19.1