nblinh63 commited on
Commit
e1179ce
·
verified ·
1 Parent(s): f3befcc

Training in progress, step 3, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "up_proj",
 
24
  "o_proj",
25
  "q_proj",
26
- "k_proj",
27
- "down_proj",
28
  "gate_proj",
29
- "v_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "v_proj",
24
+ "down_proj",
25
  "o_proj",
26
  "q_proj",
 
 
27
  "gate_proj",
28
+ "up_proj",
29
+ "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f87369f969a8727b29418baf6a1bb4e1ca43bb2c08102cdc4d024196806b875
3
  size 50503544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f9deaefb906514604bb6873bca5a2a8c97c017312d4156325e0fd14270d4af1
3
  size 50503544
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c73461fee0dda72220a2b4789b8b4c357e53616a241ad70bc6402ebfdb5ad4fe
3
  size 25986148
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5ad4b7618083a721ff31924e60ac3852e7e9960c460aafb49e5ebcdbaa98788
3
  size 25986148
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01df7b97b445c21bfe49f63c99d75eb4e03716d822d7dff7f977e3b84a2ecc05
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29425389f96c76b576bcff3abe69f0a48e741ce688d6737071ae45a2ad2d7584
3
  size 14244
last-checkpoint/trainer_state.json CHANGED
@@ -10,7 +10,7 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.00010075820549636011,
13
- "grad_norm": 6.578465480158216e+18,
14
  "learning_rate": 2e-05,
15
  "loss": 1.8654,
16
  "step": 1
@@ -18,30 +18,30 @@
18
  {
19
  "epoch": 0.00010075820549636011,
20
  "eval_loss": 1.7779617309570312,
21
- "eval_runtime": 1454.0406,
22
- "eval_samples_per_second": 2.874,
23
- "eval_steps_per_second": 1.437,
24
  "step": 1
25
  },
26
  {
27
  "epoch": 0.00020151641099272022,
28
- "grad_norm": 2.1462115130466632e+18,
29
  "learning_rate": 4e-05,
30
  "loss": 1.3359,
31
  "step": 2
32
  },
33
  {
34
  "epoch": 0.0003022746164890803,
35
- "grad_norm": 2.517414610043142e+18,
36
  "learning_rate": 6e-05,
37
- "loss": 1.3138,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0003022746164890803,
42
- "eval_loss": 1.7785661220550537,
43
- "eval_runtime": 1454.8386,
44
- "eval_samples_per_second": 2.872,
45
  "eval_steps_per_second": 1.437,
46
  "step": 3
47
  }
 
10
  "log_history": [
11
  {
12
  "epoch": 0.00010075820549636011,
13
+ "grad_norm": 8.359752932236722e+18,
14
  "learning_rate": 2e-05,
15
  "loss": 1.8654,
16
  "step": 1
 
18
  {
19
  "epoch": 0.00010075820549636011,
20
  "eval_loss": 1.7779617309570312,
21
+ "eval_runtime": 1454.9416,
22
+ "eval_samples_per_second": 2.872,
23
+ "eval_steps_per_second": 1.436,
24
  "step": 1
25
  },
26
  {
27
  "epoch": 0.00020151641099272022,
28
+ "grad_norm": 2.626370165041791e+18,
29
  "learning_rate": 4e-05,
30
  "loss": 1.3359,
31
  "step": 2
32
  },
33
  {
34
  "epoch": 0.0003022746164890803,
35
+ "grad_norm": 3.249743505089626e+18,
36
  "learning_rate": 6e-05,
37
+ "loss": 1.3042,
38
  "step": 3
39
  },
40
  {
41
  "epoch": 0.0003022746164890803,
42
+ "eval_loss": 1.7708849906921387,
43
+ "eval_runtime": 1454.6249,
44
+ "eval_samples_per_second": 2.873,
45
  "eval_steps_per_second": 1.437,
46
  "step": 3
47
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b28660bfe89cc0538651f48cd70f7a481a8052384d78f5c68e3271d050811326
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58a3eb8d82e5fbac81f07720d857f56891a89c37b161f0eab5a75ef832506307
3
  size 6776