chauhoang commited on
Commit
611153e
·
verified ·
1 Parent(s): c8a34ba

Training in progress, step 1, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -21,12 +21,12 @@
21
  "revision": null,
22
  "target_modules": [
23
  "v_proj",
24
- "down_proj",
25
- "gate_proj",
26
  "k_proj",
27
  "up_proj",
28
  "q_proj",
29
- "o_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
21
  "revision": null,
22
  "target_modules": [
23
  "v_proj",
24
+ "o_proj",
 
25
  "k_proj",
26
  "up_proj",
27
  "q_proj",
28
+ "down_proj",
29
+ "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d66e126d8cd12655145df85246f7d483721d29cc46661448e3d27b13ac63d9d
3
  size 100059752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d240ae4748e5694e07e09aeda6c8aa4d62e5ab8534b71238ee64282647d97985
3
  size 100059752
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:33f42712618c2b228b510c4a0a3b1d275897113b913d32e4cc9f509998f3b788
3
  size 51244020
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:530a3c6fab8ca4ac6cf3ebc5f5e27222b48ca4ff60fad427dc03e762226bc6c5
3
  size 51244020
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f04944917c15f265951392f0ad036c49b2e1d3572ae924179abc0eb1dc0ce1c
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bb5535922cb1f8053375cf4f0b5607ee0a1812e8ed130523ebe0c6cd11317ef
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1df0528620c07325b8faa7567e59b0c1e86a1f1ee6af1245a69c6c0463fe4e2
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae751897b8e87ff08962a91d1d3485984775a96aa89e29a1caac3d6f449228f7
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.03537944454272068,
5
- "eval_steps": 10,
6
- "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -11,127 +11,17 @@
11
  {
12
  "epoch": 0.0007075888908544136,
13
  "eval_loss": 10.552796363830566,
14
- "eval_runtime": 47.9588,
15
- "eval_samples_per_second": 12.406,
16
- "eval_steps_per_second": 6.214,
17
  "step": 1
18
- },
19
- {
20
- "epoch": 0.0035379444542720678,
21
- "grad_norm": 26.071264266967773,
22
- "learning_rate": 5e-05,
23
- "loss": 10.6088,
24
- "step": 5
25
- },
26
- {
27
- "epoch": 0.0070758889085441356,
28
- "grad_norm": 25.356914520263672,
29
- "learning_rate": 0.0001,
30
- "loss": 7.7488,
31
- "step": 10
32
- },
33
- {
34
- "epoch": 0.0070758889085441356,
35
- "eval_loss": 4.3914079666137695,
36
- "eval_runtime": 48.3524,
37
- "eval_samples_per_second": 12.305,
38
- "eval_steps_per_second": 6.163,
39
- "step": 10
40
- },
41
- {
42
- "epoch": 0.010613833362816204,
43
- "grad_norm": 28.135635375976562,
44
- "learning_rate": 9.619397662556435e-05,
45
- "loss": 2.8647,
46
- "step": 15
47
- },
48
- {
49
- "epoch": 0.014151777817088271,
50
- "grad_norm": 10.654553413391113,
51
- "learning_rate": 8.535533905932738e-05,
52
- "loss": 0.5001,
53
- "step": 20
54
- },
55
- {
56
- "epoch": 0.014151777817088271,
57
- "eval_loss": 0.41789498925209045,
58
- "eval_runtime": 48.3132,
59
- "eval_samples_per_second": 12.315,
60
- "eval_steps_per_second": 6.168,
61
- "step": 20
62
- },
63
- {
64
- "epoch": 0.01768972227136034,
65
- "grad_norm": 6.251014232635498,
66
- "learning_rate": 6.91341716182545e-05,
67
- "loss": 0.3643,
68
- "step": 25
69
- },
70
- {
71
- "epoch": 0.021227666725632408,
72
- "grad_norm": 11.83096694946289,
73
- "learning_rate": 5e-05,
74
- "loss": 0.3107,
75
- "step": 30
76
- },
77
- {
78
- "epoch": 0.021227666725632408,
79
- "eval_loss": 0.21780718863010406,
80
- "eval_runtime": 48.4603,
81
- "eval_samples_per_second": 12.278,
82
- "eval_steps_per_second": 6.149,
83
- "step": 30
84
- },
85
- {
86
- "epoch": 0.024765611179904477,
87
- "grad_norm": 0.5458418726921082,
88
- "learning_rate": 3.086582838174551e-05,
89
- "loss": 0.2974,
90
- "step": 35
91
- },
92
- {
93
- "epoch": 0.028303555634176542,
94
- "grad_norm": 7.429529190063477,
95
- "learning_rate": 1.4644660940672627e-05,
96
- "loss": 0.1783,
97
- "step": 40
98
- },
99
- {
100
- "epoch": 0.028303555634176542,
101
- "eval_loss": 0.1497228592634201,
102
- "eval_runtime": 48.4086,
103
- "eval_samples_per_second": 12.291,
104
- "eval_steps_per_second": 6.156,
105
- "step": 40
106
- },
107
- {
108
- "epoch": 0.03184150008844861,
109
- "grad_norm": 7.319284439086914,
110
- "learning_rate": 3.8060233744356633e-06,
111
- "loss": 0.1818,
112
- "step": 45
113
- },
114
- {
115
- "epoch": 0.03537944454272068,
116
- "grad_norm": 7.400360584259033,
117
- "learning_rate": 0.0,
118
- "loss": 0.3288,
119
- "step": 50
120
- },
121
- {
122
- "epoch": 0.03537944454272068,
123
- "eval_loss": 0.14456623792648315,
124
- "eval_runtime": 48.4799,
125
- "eval_samples_per_second": 12.273,
126
- "eval_steps_per_second": 6.147,
127
- "step": 50
128
  }
129
  ],
130
  "logging_steps": 5,
131
- "max_steps": 50,
132
  "num_input_tokens_seen": 0,
133
  "num_train_epochs": 1,
134
- "save_steps": 13,
135
  "stateful_callbacks": {
136
  "TrainerControl": {
137
  "args": {
@@ -144,7 +34,7 @@
144
  "attributes": {}
145
  }
146
  },
147
- "total_flos": 9555457081344000.0,
148
  "train_batch_size": 2,
149
  "trial_name": null,
150
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.0007075888908544136,
5
+ "eval_steps": 1,
6
+ "global_step": 1,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
11
  {
12
  "epoch": 0.0007075888908544136,
13
  "eval_loss": 10.552796363830566,
14
+ "eval_runtime": 47.7122,
15
+ "eval_samples_per_second": 12.471,
16
+ "eval_steps_per_second": 6.246,
17
  "step": 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  }
19
  ],
20
  "logging_steps": 5,
21
+ "max_steps": 1,
22
  "num_input_tokens_seen": 0,
23
  "num_train_epochs": 1,
24
+ "save_steps": 1,
25
  "stateful_callbacks": {
26
  "TrainerControl": {
27
  "args": {
 
34
  "attributes": {}
35
  }
36
  },
37
+ "total_flos": 191109141626880.0,
38
  "train_batch_size": 2,
39
  "trial_name": null,
40
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df7f26cf6d43d598f9b1364179ecd4079035ad9734fbe4015e87dbcee4d693bd
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e860434176efc6a708b005f903cb9015512d5a7065fbcd103d026e56154dcde9
3
  size 6776