munish0838 commited on
Commit
00e6411
1 Parent(s): da905ec

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +106 -0
README.md ADDED
@@ -0,0 +1,106 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ license: apache-2.0
5
+ datasets:
6
+ - mlabonne/orpo-dpo-mix-40k-flat
7
+ language:
8
+ - en
9
+ base_model: cognitivecomputations/dolphin-2.9.4-llama3.1-8b
10
+ pipeline_tag: text-generation
11
+ tags:
12
+ - dpo
13
+ - axolotl
14
+ library_name: transformers
15
+
16
+ ---
17
+
18
+ ![](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)
19
+
20
+ # QuantFactory/Lama-DPOlphin-8B-GGUF
21
+ This is quantized version of [CultriX/Lama-DPOlphin-8B](https://huggingface.co/CultriX/Lama-DPOlphin-8B) created using llama.cpp
22
+
23
+ # Original Model Card
24
+
25
+
26
+ ## Axolotl configuration:
27
+ ```yaml
28
+
29
+ base_model: cognitivecomputations/dolphin-2.9.4-llama3.1-8b
30
+ model_type: LlamaForCausalLM
31
+ tokenizer_type: AutoTokenizer
32
+ tokenizer:
33
+ name_or_path: "https://huggingface.co/cognitivecomputations/dolphin-2.9.4-llama3.1-8b/resolve/main/tokenizer.json"
34
+
35
+
36
+ load_in_8bit: false
37
+ load_in_4bit: true
38
+ strict: false
39
+ save_safetensors: true
40
+ bnb_4bit_quant_type: "nf4"
41
+ bnb_4bit_compute_dtype: "bf16"
42
+ bnb_4bit_use_double_quant: true
43
+
44
+ rl: dpo
45
+ chat_template: chatml
46
+ datasets:
47
+ - path: mlabonne/orpo-dpo-mix-40k-flat
48
+ split: train
49
+ type: chatml.intel
50
+
51
+ dataset_prepared_path: /workspace/axolotl/dataset-prepared
52
+ val_set_size: 0.0
53
+ output_dir: ./out
54
+
55
+ adapter: qlora
56
+ lora_model_dir:
57
+
58
+ sequence_len: 2048
59
+ sample_packing: false
60
+ pad_to_sequence_len: false
61
+
62
+ lora_r: 64
63
+ lora_alpha: 32
64
+ lora_dropout: 0.05
65
+ lora_target_linear: true
66
+ lora_fan_in_fan_out:
67
+ lora_target_modules:
68
+
69
+ wandb_project: axolotl
70
+ wandb_entity:
71
+ wandb_watch:
72
+ wandb_name:
73
+ wandb_log_model:
74
+
75
+
76
+ gradient_accumulation_steps: 4 # Reduced from 8 to 4 due to large VRAM
77
+ micro_batch_size: 2 # Increased micro-batch size to 2
78
+ num_epochs: 1
79
+ optimizer: paged_adamw_8bit
80
+ lr_scheduler: cosine
81
+ learning_rate: 5e-6
82
+ train_on_inputs: false
83
+ group_by_length: false
84
+
85
+ bf16: true # Use bf16 as it is optimal for A40 GPUs
86
+ fp16: false
87
+ tf32: true # TF32 is supported by A40 and improves performance
88
+
89
+ gradient_checkpointing: true
90
+ early_stopping_patience:
91
+ resume_from_checkpoint:
92
+ local_rank:
93
+ logging_steps: 1
94
+ xformers_attention:
95
+ flash_attention: true
96
+ warmup_steps: 100
97
+ evals_per_epoch: 0
98
+ eval_table_size:
99
+ eval_table_max_new_tokens: 128
100
+ saves_per_epoch: 1
101
+ debug:
102
+ deepspeed: deepspeed_configs/zero2.json # Enable DeepSpeed with ZeRO Stage 2
103
+ weight_decay: 0.0
104
+ special_tokens:
105
+ pad_token: <|end_of_text|>
106
+ ```