aaditya commited on
Commit
9e300ac
·
unverified ·
1 Parent(s): 3d2cd80

Adding Google's gemma Model (#1312)

Browse files
Files changed (1) hide show
  1. examples/gemma/qlora.yml +65 -0
examples/gemma/qlora.yml ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # use google/gemma-7b if you have access
2
+ base_model: mhenrichsen/gemma-7b
3
+ model_type: AutoModelForCausalLM
4
+ tokenizer_type: AutoTokenizer
5
+
6
+ load_in_8bit: false
7
+ load_in_4bit: true
8
+ strict: false
9
+
10
+ # huggingface repo
11
+ datasets:
12
+ - path: mhenrichsen/alpaca_2k_test
13
+ type: alpaca
14
+ val_set_size: 0.1
15
+ output_dir: ./out
16
+
17
+ adapter: qlora
18
+ lora_r: 32
19
+ lora_alpha: 16
20
+ lora_dropout: 0.05
21
+ lora_target_linear: true
22
+
23
+ sequence_len: 4096
24
+ sample_packing: false
25
+ pad_to_sequence_len: false
26
+
27
+ wandb_project:
28
+ wandb_entity:
29
+ wandb_watch:
30
+ wandb_name:
31
+ wandb_log_model:
32
+
33
+
34
+ gradient_accumulation_steps: 3
35
+ micro_batch_size: 2
36
+ num_epochs: 4
37
+ optimizer: adamw_bnb_8bit
38
+ lr_scheduler: cosine
39
+ learning_rate: 0.0002
40
+
41
+ train_on_inputs: false
42
+ group_by_length: false
43
+ bf16: auto
44
+ fp16:
45
+ tf32: false
46
+
47
+ gradient_checkpointing: true
48
+ early_stopping_patience:
49
+ resume_from_checkpoint:
50
+ local_rank:
51
+ logging_steps: 1
52
+ xformers_attention:
53
+ flash_attention: true
54
+
55
+ warmup_ratio: 0.1
56
+ evals_per_epoch: 4
57
+ eval_table_size:
58
+ eval_max_new_tokens: 128
59
+ saves_per_epoch: 1
60
+ debug:
61
+ deepspeed:
62
+ weight_decay: 0.0
63
+ fsdp:
64
+ fsdp_config:
65
+ special_tokens: