emre570 commited on
Commit
b817dab
1 Parent(s): f539f31

test - Updated Model

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "unsloth/gemma-1.1-7b-it-bnb-4bit",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -10,23 +10,23 @@
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
- "lora_alpha": 16,
14
- "lora_dropout": 0,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 16,
20
  "rank_pattern": {},
21
- "revision": "unsloth",
22
  "target_modules": [
23
- "gate_proj",
24
- "down_proj",
25
- "k_proj",
26
  "o_proj",
 
 
27
  "up_proj",
28
- "v_proj",
29
- "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "google/gemma-1.1-7b-it",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
10
  "layers_pattern": null,
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.0,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 32,
20
  "rank_pattern": {},
21
+ "revision": null,
22
  "target_modules": [
23
+ "v_proj",
 
 
24
  "o_proj",
25
+ "gate_proj",
26
+ "q_proj",
27
  "up_proj",
28
+ "k_proj",
29
+ "down_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9fb422ca9ce78def394728e4c71ac93f2818fffd892735a86fea01ae02245cf8
3
- size 200068512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1436fec4bc48f80ece49239fa36dbf60e970248628bf31230321a5484df168ba
3
+ size 200068904
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0da58d42d70e972774acd49a657d758c19543e002e304cb44a62f791de875959
3
+ size 200871604
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:386fcc8cc1089aade9450d86fb239ea3483f455fd2d78d8378645feecfec9d69
3
+ size 14244
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:802f82de1a07e7c596638d1365d402f6236ba5f0d2c0221ccaab7fc415f74306
3
+ size 1064
tokenizer_config.json CHANGED
@@ -1747,9 +1747,8 @@
1747
  "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
1748
  "clean_up_tokenization_spaces": false,
1749
  "eos_token": "<eos>",
1750
- "model_max_length": 8192,
1751
  "pad_token": "<pad>",
1752
- "padding_side": "right",
1753
  "sp_model_kwargs": {},
1754
  "spaces_between_special_tokens": false,
1755
  "tokenizer_class": "GemmaTokenizer",
 
1747
  "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
1748
  "clean_up_tokenization_spaces": false,
1749
  "eos_token": "<eos>",
1750
+ "model_max_length": 1000000000000000019884624838656,
1751
  "pad_token": "<pad>",
 
1752
  "sp_model_kwargs": {},
1753
  "spaces_between_special_tokens": false,
1754
  "tokenizer_class": "GemmaTokenizer",
trainer_state.json ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 12500,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.04,
13
+ "grad_norm": 2.125,
14
+ "learning_rate": 9.603841536614646e-05,
15
+ "loss": 1.8394,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.08,
20
+ "grad_norm": 2.046875,
21
+ "learning_rate": 9.203681472589036e-05,
22
+ "loss": 1.3706,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.12,
27
+ "grad_norm": 1.8984375,
28
+ "learning_rate": 8.803521408563425e-05,
29
+ "loss": 1.3084,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 0.16,
34
+ "grad_norm": 2.046875,
35
+ "learning_rate": 8.403361344537815e-05,
36
+ "loss": 1.2739,
37
+ "step": 2000
38
+ },
39
+ {
40
+ "epoch": 0.2,
41
+ "grad_norm": 1.734375,
42
+ "learning_rate": 8.003201280512204e-05,
43
+ "loss": 1.2551,
44
+ "step": 2500
45
+ },
46
+ {
47
+ "epoch": 0.24,
48
+ "grad_norm": 2.671875,
49
+ "learning_rate": 7.603041216486595e-05,
50
+ "loss": 1.2183,
51
+ "step": 3000
52
+ },
53
+ {
54
+ "epoch": 0.28,
55
+ "grad_norm": 1.7421875,
56
+ "learning_rate": 7.202881152460985e-05,
57
+ "loss": 1.1962,
58
+ "step": 3500
59
+ },
60
+ {
61
+ "epoch": 0.32,
62
+ "grad_norm": 1.8828125,
63
+ "learning_rate": 6.802721088435374e-05,
64
+ "loss": 1.1709,
65
+ "step": 4000
66
+ },
67
+ {
68
+ "epoch": 0.36,
69
+ "grad_norm": 1.9765625,
70
+ "learning_rate": 6.402561024409764e-05,
71
+ "loss": 1.1534,
72
+ "step": 4500
73
+ },
74
+ {
75
+ "epoch": 0.4,
76
+ "grad_norm": 2.078125,
77
+ "learning_rate": 6.002400960384153e-05,
78
+ "loss": 1.1418,
79
+ "step": 5000
80
+ },
81
+ {
82
+ "epoch": 0.44,
83
+ "grad_norm": 2.765625,
84
+ "learning_rate": 5.6022408963585436e-05,
85
+ "loss": 1.1375,
86
+ "step": 5500
87
+ },
88
+ {
89
+ "epoch": 0.48,
90
+ "grad_norm": 1.9453125,
91
+ "learning_rate": 5.202080832332934e-05,
92
+ "loss": 1.1271,
93
+ "step": 6000
94
+ },
95
+ {
96
+ "epoch": 0.52,
97
+ "grad_norm": 1.9453125,
98
+ "learning_rate": 4.801920768307323e-05,
99
+ "loss": 1.0945,
100
+ "step": 6500
101
+ },
102
+ {
103
+ "epoch": 0.56,
104
+ "grad_norm": 2.03125,
105
+ "learning_rate": 4.4017607042817125e-05,
106
+ "loss": 1.106,
107
+ "step": 7000
108
+ },
109
+ {
110
+ "epoch": 0.6,
111
+ "grad_norm": 2.296875,
112
+ "learning_rate": 4.001600640256102e-05,
113
+ "loss": 1.0835,
114
+ "step": 7500
115
+ },
116
+ {
117
+ "epoch": 0.64,
118
+ "grad_norm": 1.90625,
119
+ "learning_rate": 3.6014405762304925e-05,
120
+ "loss": 1.0687,
121
+ "step": 8000
122
+ },
123
+ {
124
+ "epoch": 0.68,
125
+ "grad_norm": 2.15625,
126
+ "learning_rate": 3.201280512204882e-05,
127
+ "loss": 1.0637,
128
+ "step": 8500
129
+ },
130
+ {
131
+ "epoch": 0.72,
132
+ "grad_norm": 1.8984375,
133
+ "learning_rate": 2.8011204481792718e-05,
134
+ "loss": 1.0477,
135
+ "step": 9000
136
+ },
137
+ {
138
+ "epoch": 0.76,
139
+ "grad_norm": 2.546875,
140
+ "learning_rate": 2.4009603841536614e-05,
141
+ "loss": 1.0494,
142
+ "step": 9500
143
+ },
144
+ {
145
+ "epoch": 0.8,
146
+ "grad_norm": 1.890625,
147
+ "learning_rate": 2.000800320128051e-05,
148
+ "loss": 1.0529,
149
+ "step": 10000
150
+ },
151
+ {
152
+ "epoch": 0.84,
153
+ "grad_norm": 1.6484375,
154
+ "learning_rate": 1.600640256102441e-05,
155
+ "loss": 1.0219,
156
+ "step": 10500
157
+ },
158
+ {
159
+ "epoch": 0.88,
160
+ "grad_norm": 1.8359375,
161
+ "learning_rate": 1.2004801920768307e-05,
162
+ "loss": 1.0338,
163
+ "step": 11000
164
+ },
165
+ {
166
+ "epoch": 0.92,
167
+ "grad_norm": 2.125,
168
+ "learning_rate": 8.003201280512205e-06,
169
+ "loss": 1.0454,
170
+ "step": 11500
171
+ },
172
+ {
173
+ "epoch": 0.96,
174
+ "grad_norm": 1.828125,
175
+ "learning_rate": 4.001600640256103e-06,
176
+ "loss": 1.0405,
177
+ "step": 12000
178
+ },
179
+ {
180
+ "epoch": 1.0,
181
+ "grad_norm": 1.8359375,
182
+ "learning_rate": 0.0,
183
+ "loss": 1.0363,
184
+ "step": 12500
185
+ }
186
+ ],
187
+ "logging_steps": 500,
188
+ "max_steps": 12500,
189
+ "num_input_tokens_seen": 0,
190
+ "num_train_epochs": 1,
191
+ "save_steps": 500,
192
+ "stateful_callbacks": {
193
+ "TrainerControl": {
194
+ "args": {
195
+ "should_epoch_stop": false,
196
+ "should_evaluate": false,
197
+ "should_log": false,
198
+ "should_save": true,
199
+ "should_training_stop": true
200
+ },
201
+ "attributes": {}
202
+ }
203
+ },
204
+ "total_flos": 5.596194976892191e+17,
205
+ "train_batch_size": 4,
206
+ "trial_name": null,
207
+ "trial_params": null
208
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:203c6f7d5f7372475862264ad298c4e3c9d162eadf7ba0985dd1957886cec21a
3
+ size 5112